ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2501.06248
  4. Cited By
Utility-inspired Reward Transformations Improve Reinforcement Learning Training of Language Models

Utility-inspired Reward Transformations Improve Reinforcement Learning Training of Language Models

8 January 2025
Roberto-Rafael Maura-Rivero
Chirag Nagpal
Roma Patel
Francesco Visin
ArXivPDFHTML

Papers citing "Utility-inspired Reward Transformations Improve Reinforcement Learning Training of Language Models"

1 / 1 papers shown
Title
Robust Multi-Objective Controlled Decoding of Large Language Models
Seongho Son
William Bankes
Sangwoong Yoon
Shyam Sundhar Ramesh
Xiaohang Tang
Ilija Bogunovic
34
0
0
11 Mar 2025
1