ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2406.03363
  4. Cited By
LLM-based Rewriting of Inappropriate Argumentation using Reinforcement
  Learning from Machine Feedback

LLM-based Rewriting of Inappropriate Argumentation using Reinforcement Learning from Machine Feedback

5 June 2024
Timon Ziegenbein
Gabriella Skitalinskaya
Alireza Bayat Makou
Henning Wachsmuth
    LLMAG
    KELM
ArXivPDFHTML

Papers citing "LLM-based Rewriting of Inappropriate Argumentation using Reinforcement Learning from Machine Feedback"

5 / 5 papers shown
Title
Training language models to follow instructions with human feedback
Training language models to follow instructions with human feedback
Long Ouyang
Jeff Wu
Xu Jiang
Diogo Almeida
Carroll L. Wainwright
...
Amanda Askell
Peter Welinder
Paul Christiano
Jan Leike
Ryan J. Lowe
OSLM
ALM
303
11,730
0
04 Mar 2022
Thank you BART! Rewarding Pre-Trained Models Improves Formality Style
  Transfer
Thank you BART! Rewarding Pre-Trained Models Improves Formality Style Transfer
Huiyuan Lai
Antonio Toral
Malvina Nissim
27
56
0
14 May 2021
Intrinsic Quality Assessment of Arguments
Intrinsic Quality Assessment of Arguments
Henning Wachsmuth
Till Werner
13
24
0
23 Oct 2020
Fine-Tuning Language Models from Human Preferences
Fine-Tuning Language Models from Human Preferences
Daniel M. Ziegler
Nisan Stiennon
Jeff Wu
Tom B. Brown
Alec Radford
Dario Amodei
Paul Christiano
G. Irving
ALM
275
1,561
0
18 Sep 2019
Deep Reinforcement Learning for Dialogue Generation
Deep Reinforcement Learning for Dialogue Generation
Jiwei Li
Will Monroe
Alan Ritter
Michel Galley
Jianfeng Gao
Dan Jurafsky
192
1,325
0
05 Jun 2016
1