ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2405.13541
  4. Cited By
Annotation-Efficient Language Model Alignment via Diverse and Representative Response Texts
v1v2 (latest)

Annotation-Efficient Language Model Alignment via Diverse and Representative Response Texts

22 May 2024
Yuu Jinnai
Ukyo Honda
ArXiv (abs)PDFHTMLGithub (10★)

Papers citing "Annotation-Efficient Language Model Alignment via Diverse and Representative Response Texts"

4 / 4 papers shown
Title
Sem-DPO: Mitigating Semantic Inconsistency in Preference Optimization for Prompt Engineering
Sem-DPO: Mitigating Semantic Inconsistency in Preference Optimization for Prompt Engineering
Anas Mohamed
A. Khan
Xinran Wang
Ahmad Faraz Khan
Shuwen Ge
Saman Bahzad Khan
Ayaan Ahmad
Ali Anwar
175
0
0
27 Jul 2025
The Power of Active Multi-Task Learning in Reinforcement Learning from Human Feedback
The Power of Active Multi-Task Learning in Reinforcement Learning from Human Feedback
Ruitao Chen
Liwei Wang
298
1
0
18 May 2024
Length-Controlled AlpacaEval: A Simple Way to Debias Automatic Evaluators
Length-Controlled AlpacaEval: A Simple Way to Debias Automatic Evaluators
Yann Dubois
Balázs Galambosi
Abigail Z. Jacobs
Tatsunori Hashimoto
ALM
446
585
0
06 Apr 2024
Self-Rewarding Language Models
Self-Rewarding Language Models
Weizhe Yuan
Richard Yuanzhe Pang
Kyunghyun Cho
Xian Li
Sainbayar Sukhbaatar
Jing Xu
Jason Weston
ReLMSyDaALMLRM
843
451
0
18 Jan 2024
1