Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2406.03363
Cited By
LLM-based Rewriting of Inappropriate Argumentation using Reinforcement Learning from Machine Feedback
5 June 2024
Timon Ziegenbein
Gabriella Skitalinskaya
Alireza Bayat Makou
Henning Wachsmuth
LLMAG
KELM
Re-assign community
ArXiv
PDF
HTML
Papers citing
"LLM-based Rewriting of Inappropriate Argumentation using Reinforcement Learning from Machine Feedback"
5 / 5 papers shown
Title
Training language models to follow instructions with human feedback
Long Ouyang
Jeff Wu
Xu Jiang
Diogo Almeida
Carroll L. Wainwright
...
Amanda Askell
Peter Welinder
Paul Christiano
Jan Leike
Ryan J. Lowe
OSLM
ALM
303
11,730
0
04 Mar 2022
Thank you BART! Rewarding Pre-Trained Models Improves Formality Style Transfer
Huiyuan Lai
Antonio Toral
Malvina Nissim
27
56
0
14 May 2021
Intrinsic Quality Assessment of Arguments
Henning Wachsmuth
Till Werner
13
24
0
23 Oct 2020
Fine-Tuning Language Models from Human Preferences
Daniel M. Ziegler
Nisan Stiennon
Jeff Wu
Tom B. Brown
Alec Radford
Dario Amodei
Paul Christiano
G. Irving
ALM
275
1,561
0
18 Sep 2019
Deep Reinforcement Learning for Dialogue Generation
Jiwei Li
Will Monroe
Alan Ritter
Michel Galley
Jianfeng Gao
Dan Jurafsky
192
1,325
0
05 Jun 2016
1