ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2403.01304
  4. Cited By
Improving the Validity of Automatically Generated Feedback via
  Reinforcement Learning

Improving the Validity of Automatically Generated Feedback via Reinforcement Learning

2 March 2024
Alexander Scarlatos
Digory Smith
Simon Woodhead
Andrew S. Lan
    OffRL
ArXivPDFHTML

Papers citing "Improving the Validity of Automatically Generated Feedback via Reinforcement Learning"

5 / 5 papers shown
Title
Concept-based Rubrics Improve LLM Formative Assessment and Data Synthesis
Concept-based Rubrics Improve LLM Formative Assessment and Data Synthesis
Yuchen Wei
Dennis Pearl
Matthew Beckman
Rebecca J. Passonneau
28
0
0
04 Apr 2025
Can Large Language Models Be an Alternative to Human Evaluations?
Can Large Language Models Be an Alternative to Human Evaluations?
Cheng-Han Chiang
Hung-yi Lee
ALM
LM&MA
209
568
0
03 May 2023
Large Language Models are Zero-Shot Reasoners
Large Language Models are Zero-Shot Reasoners
Takeshi Kojima
S. Gu
Machel Reid
Yutaka Matsuo
Yusuke Iwasawa
ReLM
LRM
291
4,048
0
24 May 2022
ALL-IN-ONE: Multi-Task Learning BERT models for Evaluating Peer
  Assessments
ALL-IN-ONE: Multi-Task Learning BERT models for Evaluating Peer Assessments
Qinjin Jia
Jiali Cui
Yunkai Xiao
Chengyuan Liu
Parvez Rashid
E. Gehringer
32
42
0
08 Oct 2021
Fine-Tuning Language Models from Human Preferences
Fine-Tuning Language Models from Human Preferences
Daniel M. Ziegler
Nisan Stiennon
Jeff Wu
Tom B. Brown
Alec Radford
Dario Amodei
Paul Christiano
G. Irving
ALM
275
1,583
0
18 Sep 2019
1