ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2202.05716
  4. Cited By
Choices, Risks, and Reward Reports: Charting Public Policy for
  Reinforcement Learning Systems

Choices, Risks, and Reward Reports: Charting Public Policy for Reinforcement Learning Systems

11 February 2022
T. Gilbert
Sarah Dean
Tom Zick
Nathan Lambert
    OffRL
ArXivPDFHTML

Papers citing "Choices, Risks, and Reward Reports: Charting Public Policy for Reinforcement Learning Systems"

5 / 5 papers shown
Title
The Alignment Ceiling: Objective Mismatch in Reinforcement Learning from
  Human Feedback
The Alignment Ceiling: Objective Mismatch in Reinforcement Learning from Human Feedback
Nathan Lambert
Roberto Calandra
ALM
13
31
0
31 Oct 2023
Designing Fiduciary Artificial Intelligence
Designing Fiduciary Artificial Intelligence
Sebastian Benthall
David Shekman
43
3
0
27 Jul 2023
Dynamic Documentation for AI Systems
Dynamic Documentation for AI Systems
S. Mehta
Anderson Rogers
T. Gilbert
13
0
0
20 Mar 2023
Actionable Guidance for High-Consequence AI Risk Management: Towards
  Standards Addressing AI Catastrophic Risks
Actionable Guidance for High-Consequence AI Risk Management: Towards Standards Addressing AI Catastrophic Risks
Anthony M. Barrett
Dan Hendrycks
Jessica Newman
Brandie Nonnecke
SILM
10
11
0
17 Jun 2022
Reward Reports for Reinforcement Learning
Reward Reports for Reinforcement Learning
T. Gilbert
Nathan Lambert
Sarah Dean
Tom Zick
Aaron J. Snoswell
32
33
0
22 Apr 2022
1