ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2503.10093
  4. Cited By

Representation-based Reward Modeling for Efficient Safety Alignment of Large Language Model

13 March 2025
Qiyuan Deng
X. Bai
Kehai Chen
Yaowei Wang
Liqiang Nie
Min Zhang
    OffRL
ArXivPDFHTML

Papers citing "Representation-based Reward Modeling for Efficient Safety Alignment of Large Language Model"

Title
No papers