Communities
Connect sessions
AI calendar
Organizations
Join Slack
Contact Sales
Search
Open menu
Home
Papers
2503.00539
Cited By
Distributionally Robust Reinforcement Learning with Human Feedback
1 March 2025
Debmalya Mandal
Paulius Sasnauskas
Goran Radanović
Re-assign community
ArXiv (abs)
PDF
HTML
Github
Papers citing
"Distributionally Robust Reinforcement Learning with Human Feedback"
6 / 6 papers shown
General Intelligence-based Fragmentation (GIF): A framework for peak-labeled spectra simulation
Margaret R. Martin
Soha Hassoun
128
0
0
11 Nov 2025
Lightweight Robust Direct Preference Optimization
Cheol Woo Kim
Shresth Verma
Mauricio Tec
Milind Tambe
194
2
0
27 Oct 2025
Understanding the Performance Gap in Preference Learning: A Dichotomy of RLHF and DPO
Ruizhe Shi
Minhak Song
Runlong Zhou
Zihan Zhang
Maryam Fazel
S. S. Du
494
8
0
26 May 2025
Robust Reinforcement Learning from Human Feedback for Large Language Models Fine-Tuning
Kai Ye
Hongyi Zhou
Jin Zhu
Francesco Quinzan
C. Shi
722
13
0
03 Apr 2025
Uncertainty-aware Reward Model: Teaching Reward Models to Know What is Unknown
Xingzhou Lou
Dong Yan
Wei Shen
Yuzi Yan
Jian Xie
Junge Zhang
460
57
0
01 Oct 2024
Right Now, Wrong Then: Non-Stationary Direct Preference Optimization under Preference Drift
Seongho Son
William Bankes
Sayak Ray Chowdhury
Brooks Paige
Ilija Bogunovic
496
9
0
26 Jul 2024
1
Page 1 of 1