321
v1v2 (latest)

RePO: Understanding Preference Learning Through ReLU-Based Optimization

Main:8 Pages
10 Figures
Bibliography:6 Pages
8 Tables
Appendix:15 Pages
Abstract

Aligning large language models (LLMs) with human preferences is critical for real-world deployment, yet existing methods like RLHF face computational and stability challenges. While DPO establishes an offline paradigm with single hyperparameter β\beta, subsequent methods like SimPO reintroduce complexity through dual parameters (β\beta, γ\gamma). We propose {ReLU-based Preference Optimization (RePO)}, a streamlined algorithm that eliminates β\beta via two advances: (1) retaining SimPO's reference-free margins but removing β\beta through gradient analysis, and (2) adopting a ReLU-based max-margin loss that naturally filters trivial pairs. Theoretically, RePO is characterized as SimPO's limiting case (β\beta \to \infty), where the logistic weighting collapses to binary thresholding, forming a convex envelope of the 0-1 loss. Empirical results on AlpacaEval 2 and Arena-Hard show that RePO outperforms DPO and SimPO across multiple base models, requiring only one hyperparameter to tune.

View on arXiv
Comments on this paper