Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2403.14683
Cited By
A Moral Imperative: The Need for Continual Superalignment of Large Language Models
13 March 2024
Gokul Puthumanaillam
Manav Vora
Pranay Thangeda
Melkior Ornik
Re-assign community
ArXiv
PDF
HTML
Papers citing
"A Moral Imperative: The Need for Continual Superalignment of Large Language Models"
5 / 5 papers shown
Title
Your Weak LLM is Secretly a Strong Teacher for Alignment
Leitian Tao
Yixuan Li
79
5
0
13 Sep 2024
Hallucination Detection in Foundation Models for Decision-Making: A Flexible Definition and Review of the State of the Art
Neeloy Chakraborty
Melkior Ornik
Katherine Driggs-Campbell
LRM
43
9
0
25 Mar 2024
Large Language Models for Robotics: A Survey
Fanlong Zeng
Wensheng Gan
Yongheng Wang
Ning Liu
Philip S. Yu
LM&Ro
106
121
0
13 Nov 2023
Self-RAG: Learning to Retrieve, Generate, and Critique through Self-Reflection
Akari Asai
Zeqiu Wu
Yizhong Wang
Avirup Sil
Hannaneh Hajishirzi
RALM
138
600
0
17 Oct 2023
Towards Continual Knowledge Learning of Language Models
Joel Jang
Seonghyeon Ye
Sohee Yang
Joongbo Shin
Janghoon Han
Gyeonghun Kim
Stanley Jungkyu Choi
Minjoon Seo
CLL
KELM
216
122
0
07 Oct 2021
1