Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2406.00832
Cited By
BoNBoN Alignment for Large Language Models and the Sweetness of Best-of-n Sampling
2 June 2024
Lin Gui
Cristina Garbacea
Victor Veitch
BDL
LM&MA
Re-assign community
ArXiv
PDF
HTML
Papers citing
"BoNBoN Alignment for Large Language Models and the Sweetness of Best-of-n Sampling"
5 / 5 papers shown
Title
Unintentional Unalignment: Likelihood Displacement in Direct Preference Optimization
Noam Razin
Sadhika Malladi
Adithya Bhaskar
Danqi Chen
Sanjeev Arora
Boris Hanin
39
10
0
11 Oct 2024
Asymptotics of Language Model Alignment
Joy Qiping Yang
Salman Salamatian
Ziteng Sun
A. Suresh
Ahmad Beirami
23
12
0
02 Apr 2024
KTO: Model Alignment as Prospect Theoretic Optimization
Kawin Ethayarajh
Winnie Xu
Niklas Muennighoff
Dan Jurafsky
Douwe Kiela
125
147
0
02 Feb 2024
Training language models to follow instructions with human feedback
Long Ouyang
Jeff Wu
Xu Jiang
Diogo Almeida
Carroll L. Wainwright
...
Amanda Askell
Peter Welinder
Paul Christiano
Jan Leike
Ryan J. Lowe
OSLM
ALM
270
8,441
0
04 Mar 2022
Fine-Tuning Language Models from Human Preferences
Daniel M. Ziegler
Nisan Stiennon
Jeff Wu
Tom B. Brown
Alec Radford
Dario Amodei
Paul Christiano
G. Irving
ALM
245
1,151
0
18 Sep 2019
1