ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2305.13534
  4. Cited By
How Language Model Hallucinations Can Snowball

How Language Model Hallucinations Can Snowball

International Conference on Machine Learning (ICML), 2023
22 May 2023
Muru Zhang
Ofir Press
William Merrill
Alisa Liu
Noah A. Smith
    HILMLRM
ArXiv (abs)PDFHTMLHuggingFace (3 upvotes)

Papers citing "How Language Model Hallucinations Can Snowball"

50 / 125 papers shown
MAMM-Refine: A Recipe for Improving Faithfulness in Generation with Multi-Agent Collaboration
MAMM-Refine: A Recipe for Improving Faithfulness in Generation with Multi-Agent CollaborationNorth American Chapter of the Association for Computational Linguistics (NAACL), 2025
David Wan
Justin Chih-Yao Chen
Elias Stengel-Eskin
Joey Tianyi Zhou
LLMAGLRM
275
5
0
19 Mar 2025
Where do Large Vision-Language Models Look at when Answering Questions?
Where do Large Vision-Language Models Look at when Answering Questions?
X. Xing
Chia-Wen Kuo
Li Fuxin
Yulei Niu
Fan Chen
Ming Li
Ying Wu
Longyin Wen
Sijie Zhu
LRM
284
6
0
18 Mar 2025
DatawiseAgent: A Notebook-Centric LLM Agent Framework for Adaptive and Robust Data Science Automation
DatawiseAgent: A Notebook-Centric LLM Agent Framework for Adaptive and Robust Data Science Automation
Ziming You
Yumiao Zhang
Dexuan Xu
Yiwei Lou
Yandong Yan
Wei Wang
H. Zhang
Yu Huang
LLMAG
284
4
0
10 Mar 2025
Can LLMs Explain Themselves Counterfactually?
Can LLMs Explain Themselves Counterfactually?
Zahra Dehghanighobadi
Asja Fischer
Muhammad Bilal Zafar
LRM
408
2
0
25 Feb 2025
GraphCheck: Breaking Long-Term Text Barriers with Extracted Knowledge Graph-Powered Fact-Checking
GraphCheck: Breaking Long-Term Text Barriers with Extracted Knowledge Graph-Powered Fact-CheckingAnnual Meeting of the Association for Computational Linguistics (ACL), 2025
Yingjian Chen
Haoran Liu
Yinhong Liu
Rui Yang
Han Yuan
...
Pengyuan Zhou
Peng Yuan Zhou
Qingyu Chen
James Caverlee
Irene Li
HILM
570
7
0
23 Feb 2025
Preventing Rogue Agents Improves Multi-Agent Collaboration
Preventing Rogue Agents Improves Multi-Agent Collaboration
Ohav Barbi
Ori Yoran
Mor Geva
327
6
0
09 Feb 2025
ComparisonQA: Evaluating Factuality Robustness of LLMs Through Knowledge Frequency Control and Uncertainty
ComparisonQA: Evaluating Factuality Robustness of LLMs Through Knowledge Frequency Control and UncertaintyAnnual Meeting of the Association for Computational Linguistics (ACL), 2024
Qing Zong
Zhaoxiang Wang
Tianshi Zheng
Xiyu Ren
Yangqiu Song
388
11
0
28 Dec 2024
The Potential of LLMs in Medical Education: Generating Questions and Answers for Qualification Exams
The Potential of LLMs in Medical Education: Generating Questions and Answers for Qualification Exams
Yunqi Zhu
Wen Tang
Ying Sun
Xuebing Yang
Liyang Dou
Yifan Gu
Yuanyuan Wu
Wensheng Zhang
Ying Sun
Xuebing Yang
LM&MAELM
317
5
0
31 Oct 2024
Retrieval-Augmented Generation with Estimation of Source Reliability
Retrieval-Augmented Generation with Estimation of Source Reliability
Jeongyeon Hwang
Junyoung Park
Hyejin Park
Dongwoo Kim
Sangdon Park
Jungseul Ok
RALM
463
4
0
30 Oct 2024
LabSafety Bench: Benchmarking LLMs on Safety Issues in Scientific Labs
LabSafety Bench: Benchmarking LLMs on Safety Issues in Scientific Labs
Yujun Zhou
Jingdong Yang
Yue Huang
Kehan Guo
Zoe Emory
...
Tian Gao
Werner Geyer
Nuno Moniz
Nitesh Chawla
Xiangliang Zhang
466
12
0
18 Oct 2024
QSpec: Speculative Decoding with Complementary Quantization Schemes
QSpec: Speculative Decoding with Complementary Quantization Schemes
Juntao Zhao
Wenhao Lu
Sheng Wang
Lingpeng Kong
Chuan Wu
MQ
436
11
0
15 Oct 2024
A Survey on the Honesty of Large Language Models
A Survey on the Honesty of Large Language Models
Siheng Li
Cheng Yang
Taiqiang Wu
Chufan Shi
Yuji Zhang
...
Jie Zhou
Yujiu Yang
Ngai Wong
Xixin Wu
Wai Lam
HILM
293
18
0
27 Sep 2024
Multilevel Interpretability Of Artificial Neural Networks: Leveraging
  Framework And Methods From Neuroscience
Multilevel Interpretability Of Artificial Neural Networks: Leveraging Framework And Methods From Neuroscience
Zhonghao He
Jascha Achterberg
Katie Collins
Kevin K. Nejad
Danyal Akarca
...
Chole Li
Kai J. Sandbrink
Stephen Casper
Anna Ivanova
Grace W. Lindsay
AI4CE
322
6
0
22 Aug 2024
Visual Agents as Fast and Slow Thinkers
Visual Agents as Fast and Slow ThinkersInternational Conference on Learning Representations (ICLR), 2024
Guangyan Sun
Haoyang Ling
Zhenting Wang
Cheng-Long Wang
Siqi Ma
Qifan Wang
Ying Nian Wu
Ying Nian Wu
Dongfang Liu
Dongfang Liu
LLMAGLRM
529
43
0
16 Aug 2024
ExoViP: Step-by-step Verification and Exploration with Exoskeleton
  Modules for Compositional Visual Reasoning
ExoViP: Step-by-step Verification and Exploration with Exoskeleton Modules for Compositional Visual Reasoning
Yanjie Wang
Alan Yuille
Zhuowan Li
Zilong Zheng
LRM
312
6
0
05 Aug 2024
Self-Introspective Decoding: Alleviating Hallucinations for Large Vision-Language Models
Self-Introspective Decoding: Alleviating Hallucinations for Large Vision-Language ModelsInternational Conference on Learning Representations (ICLR), 2024
Fushuo Huo
Wenchao Xu
Zhong Zhang
Yining Qi
Zhicheng Chen
Peilin Zhao
VLMMLLM
579
58
0
04 Aug 2024
Social and Ethical Risks Posed by General-Purpose LLMs for Settling Newcomers in Canada
Social and Ethical Risks Posed by General-Purpose LLMs for Settling Newcomers in Canada
I. Nejadgholi
Maryam Molamohammadi
Samir Bakhtawar
380
0
0
15 Jul 2024
Merge, Ensemble, and Cooperate! A Survey on Collaborative Strategies in
  the Era of Large Language Models
Merge, Ensemble, and Cooperate! A Survey on Collaborative Strategies in the Era of Large Language Models
Jinliang Lu
Ziliang Pang
Min Xiao
Yaochen Zhu
Rui Xia
Jiajun Zhang
MoMe
384
48
0
08 Jul 2024
From Loops to Oops: Fallback Behaviors of Language Models Under Uncertainty
From Loops to Oops: Fallback Behaviors of Language Models Under Uncertainty
Maor Ivgi
Ori Yoran
Jonathan Berant
Mor Geva
HILM
425
9
0
08 Jul 2024
Predicting vs. Acting: A Trade-off Between World Modeling & Agent
  Modeling
Predicting vs. Acting: A Trade-off Between World Modeling & Agent Modeling
Margaret Li
Weijia Shi
Artidoro Pagnoni
Peter West
Ari Holtzman
232
16
0
02 Jul 2024
Learning to Refine with Fine-Grained Natural Language Feedback
Learning to Refine with Fine-Grained Natural Language Feedback
Manya Wadhwa
Xinyu Zhao
Junyi Jessy Li
Greg Durrett
532
25
0
02 Jul 2024
First Heuristic Then Rational: Dynamic Use of Heuristics in Language
  Model Reasoning
First Heuristic Then Rational: Dynamic Use of Heuristics in Language Model Reasoning
Yoichi Aoki
Keito Kudo
Tatsuki Kuribayashi
Shusaku Sone
Masaya Taniguchi
Keisuke Sakaguchi
Kentaro Inui
LRM
333
1
0
23 Jun 2024
Chain-of-Probe: Examining the Necessity and Accuracy of CoT Step-by-Step
Chain-of-Probe: Examining the Necessity and Accuracy of CoT Step-by-Step
Zezhong Wang
Xingshan Zeng
Weiwen Liu
Yufei Wang
Liangyou Li
Yasheng Wang
Lifeng Shang
Xin Jiang
Qun Liu
Kam-Fai Wong
LRM
255
5
0
23 Jun 2024
A Probabilistic Framework for LLM Hallucination Detection via Belief Tree Propagation
A Probabilistic Framework for LLM Hallucination Detection via Belief Tree Propagation
Bairu Hou
Yang Zhang
Jacob Andreas
Shiyu Chang
299
18
0
11 Jun 2024
ANAH: Analytical Annotation of Hallucinations in Large Language Models
ANAH: Analytical Annotation of Hallucinations in Large Language Models
Ziwei Ji
Yuzhe Gu
Wenwei Zhang
Chengqi Lyu
Dahua Lin
Kai-xiang Chen
HILM
203
8
0
30 May 2024
Improve Student's Reasoning Generalizability through Cascading
  Decomposed CoTs Distillation
Improve Student's Reasoning Generalizability through Cascading Decomposed CoTs Distillation
Chengwei Dai
Kun Li
Wei Zhou
Song Hu
LRM
235
13
0
30 May 2024
OpenFactCheck: Building, Benchmarking Customized Fact-Checking Systems and Evaluating the Factuality of Claims and LLMs
OpenFactCheck: Building, Benchmarking Customized Fact-Checking Systems and Evaluating the Factuality of Claims and LLMs
Yuxia Wang
Minghan Wang
Hasan Iqbal
Georgi Georgiev
Fauzan Farooqui
Preslav Nakov
HILM
415
36
0
09 May 2024
The Landscape of Emerging AI Agent Architectures for Reasoning,
  Planning, and Tool Calling: A Survey
The Landscape of Emerging AI Agent Architectures for Reasoning, Planning, and Tool Calling: A Survey
Tula Masterman
Sandi Besen
Mason Sawtell
Alex Chao
LM&RoLLMAG
330
136
0
17 Apr 2024
Self-playing Adversarial Language Game Enhances LLM Reasoning
Self-playing Adversarial Language Game Enhances LLM Reasoning
Pengyu Cheng
Tianhao Hu
Han Xu
Zhisong Zhang
Yong Dai
Lei Han
Nan Du
Nan Du
Xiaolong Li
SyDaLRMReLM
492
72
0
16 Apr 2024
Automating Research Synthesis with Domain-Specific Large Language Model
  Fine-Tuning
Automating Research Synthesis with Domain-Specific Large Language Model Fine-Tuning
Teo Susnjak
Peter Hwang
N. Reyes
A. Barczak
Timothy R. McIntosh
Surangika Ranathunga
231
56
0
08 Apr 2024
Hallucination Detection in Foundation Models for Decision-Making: A Flexible Definition and Review of the State of the Art
Hallucination Detection in Foundation Models for Decision-Making: A Flexible Definition and Review of the State of the Art
Neeloy Chakraborty
Melkior Ornik
Katherine Driggs-Campbell
LRM
433
27
0
25 Mar 2024
FIT-RAG: Black-Box RAG with Factual Information and Token Reduction
FIT-RAG: Black-Box RAG with Factual Information and Token Reduction
Yuren Mao
Xuemei Dong
Wenyi Xu
Yunjun Gao
Bin Wei
Ying Zhang
193
18
0
21 Mar 2024
ERBench: An Entity-Relationship based Automatically Verifiable
  Hallucination Benchmark for Large Language Models
ERBench: An Entity-Relationship based Automatically Verifiable Hallucination Benchmark for Large Language ModelsNeural Information Processing Systems (NeurIPS), 2024
Jio Oh
Soyeon Kim
Junseok Seo
Yongfeng Zhang
Ruochen Xu
Xing Xie
Steven Euijong Whang
217
11
0
08 Mar 2024
SPUQ: Perturbation-Based Uncertainty Quantification for Large Language
  Models
SPUQ: Perturbation-Based Uncertainty Quantification for Large Language Models
Xiang Gao
Jiaxin Zhang
Lalla Mouatadid
Kamalika Das
246
24
0
04 Mar 2024
Making Reasoning Matter: Measuring and Improving Faithfulness of
  Chain-of-Thought Reasoning
Making Reasoning Matter: Measuring and Improving Faithfulness of Chain-of-Thought Reasoning
Debjit Paul
Robert West
Antoine Bosselut
Boi Faltings
ReLMLRM
420
77
0
21 Feb 2024
Rowen: Adaptive Retrieval-Augmented Generation for Hallucination Mitigation in LLMs
Rowen: Adaptive Retrieval-Augmented Generation for Hallucination Mitigation in LLMs
Hanxing Ding
Liang Pang
Zihao Wei
Huawei Shen
Xueqi Cheng
HILMRALM
432
26
0
16 Feb 2024
A Chain-of-Thought Is as Strong as Its Weakest Link: A Benchmark for
  Verifiers of Reasoning Chains
A Chain-of-Thought Is as Strong as Its Weakest Link: A Benchmark for Verifiers of Reasoning Chains
Alon Jacovi
Yonatan Bitton
Bernd Bohnet
Jonathan Herzig
Or Honovich
Michael Tseng
Michael Collins
Roee Aharoni
Mor Geva
LRM
430
43
0
01 Feb 2024
Deductive Beam Search: Decoding Deducible Rationale for Chain-of-Thought
  Reasoning
Deductive Beam Search: Decoding Deducible Rationale for Chain-of-Thought Reasoning
Tinghui Zhu
Kai Zhang
Jian Xie
Yu-Chuan Su
LRM
320
23
0
31 Jan 2024
Generative AI in EU Law: Liability, Privacy, Intellectual Property, and
  Cybersecurity
Generative AI in EU Law: Liability, Privacy, Intellectual Property, and CybersecuritySocial Science Research Network (SSRN), 2024
Claudio Novelli
F. Casolari
Philipp Hacker
Giorgio Spedicato
Luciano Floridi
AILawSILM
440
97
0
14 Jan 2024
GPT-4V(ision) is a Human-Aligned Evaluator for Text-to-3D Generation
GPT-4V(ision) is a Human-Aligned Evaluator for Text-to-3D GenerationComputer Vision and Pattern Recognition (CVPR), 2024
Tong Wu
Guandao Yang
Zhibing Li
Kai Zhang
Ziwei Liu
Leonidas Guibas
Dahua Lin
Gordon Wetzstein
EGVMVGen
397
138
0
08 Jan 2024
DCR-Consistency: Divide-Conquer-Reasoning for Consistency Evaluation and
  Improvement of Large Language Models
DCR-Consistency: Divide-Conquer-Reasoning for Consistency Evaluation and Improvement of Large Language Models
Wendi Cui
Jiaxin Zhang
Zhuohang Li
Lopez Damien
Kamalika Das
Sricharan Kumar
Kumar Sricharan
188
2
0
04 Jan 2024
The Persuasive Power of Large Language Models
The Persuasive Power of Large Language Models
Simon Martin Breum
Daniel Vaedele Egdal
Victor Gram Mortensen
Anders Giovanni Møller
L. Aiello
AI4CE
204
53
0
24 Dec 2023
LLM-SQL-Solver: Can LLMs Determine SQL Equivalence?
LLM-SQL-Solver: Can LLMs Determine SQL Equivalence?
Fuheng Zhao
Lawrence Lim
Ishtiyaque Ahmad
D. Agrawal
A. El Abbadi
Amr El Abbadi
536
19
0
16 Dec 2023
Making Large Language Models Better Knowledge Miners for Online Marketing with Progressive Prompting Augmentation
Making Large Language Models Better Knowledge Miners for Online Marketing with Progressive Prompting Augmentation
Chunjing Gan
Dan Yang
Binbin Hu
Ziqi Liu
Yue Shen
Qing Cui
Jinjie Gu
Jun Zhou
Guannan Zhang
343
8
0
08 Dec 2023
OPERA: Alleviating Hallucination in Multi-Modal Large Language Models
  via Over-Trust Penalty and Retrospection-Allocation
OPERA: Alleviating Hallucination in Multi-Modal Large Language Models via Over-Trust Penalty and Retrospection-AllocationComputer Vision and Pattern Recognition (CVPR), 2023
Qidong Huang
Xiao-wen Dong
Pan Zhang
Sijin Yu
Conghui He
Yuan Liu
Dahua Lin
Weiming Zhang
Neng H. Yu
MLLM
463
360
0
29 Nov 2023
Calibrated Language Models Must Hallucinate
Calibrated Language Models Must HallucinateSymposium on the Theory of Computing (STOC), 2023
Adam Tauman Kalai
Santosh Vempala
HILM
399
129
0
24 Nov 2023
Ever: Mitigating Hallucination in Large Language Models through
  Real-Time Verification and Rectification
Ever: Mitigating Hallucination in Large Language Models through Real-Time Verification and Rectification
Haoqiang Kang
Juntong Ni
Huaxiu Yao
HILMLRM
396
46
0
15 Nov 2023
Can Knowledge Graphs Reduce Hallucinations in LLMs? : A Survey
Can Knowledge Graphs Reduce Hallucinations in LLMs? : A SurveyNorth American Chapter of the Association for Computational Linguistics (NAACL), 2023
Garima Agrawal
Tharindu Kumarage
Zeyad Alghami
Huanmin Liu
312
145
0
14 Nov 2023
A Survey on Hallucination in Large Language Models: Principles,
  Taxonomy, Challenges, and Open Questions
A Survey on Hallucination in Large Language Models: Principles, Taxonomy, Challenges, and Open Questions
Lei Huang
Weijiang Yu
Weitao Ma
Weihong Zhong
Zhangyin Feng
...
Qianglong Chen
Weihua Peng
Xiaocheng Feng
Bing Qin
Ting Liu
LRMHILM
437
1,930
0
09 Nov 2023
In-Context Learning Dynamics with Random Binary Sequences
In-Context Learning Dynamics with Random Binary SequencesInternational Conference on Learning Representations (ICLR), 2023
Eric J. Bigelow
Ekdeep Singh Lubana
Robert P. Dick
Hidenori Tanaka
T. Ullman
425
12
0
26 Oct 2023
Previous
123
Next