ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2503.11751
  4. Cited By
reWordBench: Benchmarking and Improving the Robustness of Reward Models with Transformed Inputs

reWordBench: Benchmarking and Improving the Robustness of Reward Models with Transformed Inputs

14 March 2025
Zhaofeng Wu
Michihiro Yasunaga
Andrew Cohen
Yoon Kim
Asli Celikyilmaz
Marjan Ghazvininejad
ArXivPDFHTML

Papers citing "reWordBench: Benchmarking and Improving the Robustness of Reward Models with Transformed Inputs"

1 / 1 papers shown
Title
Adversarial Training of Reward Models
Adversarial Training of Reward Models
Alexander Bukharin
Haifeng Qian
Shengyang Sun
Adithya Renduchintala
Soumye Singhal
Z. Wang
Oleksii Kuchaiev
Olivier Delalleau
T. Zhao
AAML
27
0
0
08 Apr 2025
1