Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2406.15178
Cited By
Hybrid Alignment Training for Large Language Models
21 June 2024
Chenglong Wang
Hang Zhou
Kaiyan Chang
Bei Li
Yongyu Mu
Tong Xiao
Tongran Liu
Jingbo Zhu
Re-assign community
ArXiv
PDF
HTML
Papers citing
"Hybrid Alignment Training for Large Language Models"
2 / 2 papers shown
Title
Cascade Reward Sampling for Efficient Decoding-Time Alignment
Bolian Li
Yifan Wang
A. Grama
Ruqi Zhang
Ruqi Zhang
AI4TS
44
8
0
24 Jun 2024
Training language models to follow instructions with human feedback
Long Ouyang
Jeff Wu
Xu Jiang
Diogo Almeida
Carroll L. Wainwright
...
Amanda Askell
Peter Welinder
Paul Christiano
Jan Leike
Ryan J. Lowe
OSLM
ALM
301
11,730
0
04 Mar 2022
1