Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2403.08309
Cited By
HRLAIF: Improvements in Helpfulness and Harmlessness in Open-domain Reinforcement Learning From AI Feedback
13 March 2024
Ang Li
Qiugen Xiao
Peng Cao
Jian Tang
Yi Yuan
Zijie Zhao
Xiaoyuan Chen
Liang Zhang
Xiangyang Li
Kaitong Yang
Weidong Guo
Yukang Gan
Jeffrey Xu Yu
D. Wang
Ying Shan
VLM
ALM
Re-assign community
ArXiv
PDF
HTML
Papers citing
"HRLAIF: Improvements in Helpfulness and Harmlessness in Open-domain Reinforcement Learning From AI Feedback"
4 / 4 papers shown
Title
An Overview and Discussion on Using Large Language Models for Implementation Generation of Solutions to Open-Ended Problems
Hashmath Shaik
Alex Doboli
OffRL
ELM
63
0
0
31 Dec 2024
Seeing Eye to AI: Human Alignment via Gaze-Based Response Rewards for Large Language Models
Angela Lopez-Cardona
Carlos Segura
Alexandros Karatzoglou
Sergi Abadal
Ioannis Arapakis
ALM
48
2
0
02 Oct 2024
An Empirical Study of Instruction-tuning Large Language Models in Chinese
Q. Si
Tong Wang
Zheng Lin
Xu Zhang
Yanan Cao
Weiping Wang
ALM
54
16
0
11 Oct 2023
Training language models to follow instructions with human feedback
Long Ouyang
Jeff Wu
Xu Jiang
Diogo Almeida
Carroll L. Wainwright
...
Amanda Askell
Peter Welinder
Paul Christiano
Jan Leike
Ryan J. Lowe
OSLM
ALM
301
11,730
0
04 Mar 2022
1