ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2404.18930
  4. Cited By
Hallucination of Multimodal Large Language Models: A Survey
v1v2 (latest)

Hallucination of Multimodal Large Language Models: A Survey

29 April 2024
Zechen Bai
Pichao Wang
Tianjun Xiao
Tong He
Zongbo Han
Zheng Zhang
Mike Zheng Shou
    VLMLRM
ArXiv (abs)PDFHTML

Papers citing "Hallucination of Multimodal Large Language Models: A Survey"

50 / 334 papers shown
Measuring and Enhancing Trustworthiness of LLMs in RAG through Grounded Attributions and Learning to Refuse
Measuring and Enhancing Trustworthiness of LLMs in RAG through Grounded Attributions and Learning to RefuseInternational Conference on Learning Representations (ICLR), 2024
Maojia Song
Shang Hong Sim
Rishabh Bhardwaj
Hai Leong Chieu
Navonil Majumder
Soujanya Poria
562
28
0
17 Sep 2024
Surveying the MLLM Landscape: A Meta-Review of Current Surveys
Surveying the MLLM Landscape: A Meta-Review of Current Surveys
Ming Li
Keyu Chen
Ziqian Bi
Ming Liu
Xinyuan Song
...
Jinlang Wang
Sen Zhang
Xuanhe Pan
Jiawei Xu
Pohsun Feng
OffRL
277
12
0
17 Sep 2024
When Context Leads but Parametric Memory Follows in Large Language
  Models
When Context Leads but Parametric Memory Follows in Large Language ModelsConference on Empirical Methods in Natural Language Processing (EMNLP), 2024
Yufei Tao
Adam Hiatt
Erik Haake
Antonie J. Jetter
Ameeta Agrawal
KELM
355
6
0
13 Sep 2024
LaMsS: When Large Language Models Meet Self-Skepticism
LaMsS: When Large Language Models Meet Self-Skepticism
Yetao Wu
Yihong Wang
Teng Chen
Chenxi Liu
Ningyuan Xi
Qingqing Gu
Hongyang Lei
HILM
145
1
0
10 Sep 2024
Visual Prompting in Multimodal Large Language Models: A Survey
Visual Prompting in Multimodal Large Language Models: A Survey
Junda Wu
Zhehao Zhang
Yu Xia
Xintong Li
Zhaoyang Xia
...
Subrata Mitra
Dimitris N. Metaxas
Lina Yao
Jingbo Shang
Julian McAuley
VLMLRM
314
42
0
05 Sep 2024
Understanding Multimodal Hallucination with Parameter-Free
  Representation Alignment
Understanding Multimodal Hallucination with Parameter-Free Representation Alignment
Yueqian Wang
Jianxin Liang
Yuxuan Wang
Huishuai Zhang
Dongyan Zhao
237
2
0
02 Sep 2024
Look, Compare, Decide: Alleviating Hallucination in Large
  Vision-Language Models via Multi-View Multi-Path Reasoning
Look, Compare, Decide: Alleviating Hallucination in Large Vision-Language Models via Multi-View Multi-Path ReasoningInternational Conference on Computational Linguistics (COLING), 2024
Xiaoye Qu
Jiashuo Sun
Wei Wei
Yu Cheng
MLLMLRM
277
21
0
30 Aug 2024
Has Multimodal Learning Delivered Universal Intelligence in Healthcare?
  A Comprehensive Survey
Has Multimodal Learning Delivered Universal Intelligence in Healthcare? A Comprehensive SurveyInformation Fusion (Inf. Fusion), 2024
Qika Lin
Yifan Zhu
Xin Mei
Ling Huang
Jingying Ma
Kai He
Zhen Peng
Xiaoshi Zhong
Mengling Feng
292
61
0
23 Aug 2024
Show-o: One Single Transformer to Unify Multimodal Understanding and
  Generation
Show-o: One Single Transformer to Unify Multimodal Understanding and GenerationInternational Conference on Learning Representations (ICLR), 2024
Jinheng Xie
Weijia Mao
Zechen Bai
David Junhao Zhang
Weihao Wang
Kevin Qinghong Lin
Yuchao Gu
Zhijie Chen
Zhenheng Yang
Mike Zheng Shou
401
441
0
22 Aug 2024
Transfusion: Predict the Next Token and Diffuse Images with One
  Multi-Modal Model
Transfusion: Predict the Next Token and Diffuse Images with One Multi-Modal Model
Chunting Zhou
Lili Yu
Arun Babu
Kushal Tirumala
Michihiro Yasunaga
Leonid Shamis
Jacob Kahn
Xuezhe Ma
Luke Zettlemoyer
Omer Levy
DiffM
265
294
0
20 Aug 2024
CLIP-DPO: Vision-Language Models as a Source of Preference for Fixing
  Hallucinations in LVLMs
CLIP-DPO: Vision-Language Models as a Source of Preference for Fixing Hallucinations in LVLMsEuropean Conference on Computer Vision (ECCV), 2024
Yassine Ouali
Adrian Bulat
Brais Martínez
Georgios Tzimiropoulos
VLMMLLM
298
42
0
19 Aug 2024
CoVLA: Comprehensive Vision-Language-Action Dataset for Autonomous Driving
CoVLA: Comprehensive Vision-Language-Action Dataset for Autonomous DrivingIEEE Workshop/Winter Conference on Applications of Computer Vision (WACV), 2024
Hidehisa Arai
Keita Miwa
Kento Sasaki
Yu Yamaguchi
Kohei Watanabe
Shunsuke Aoki
Issei Yamamoto
350
48
0
19 Aug 2024
Reefknot: A Comprehensive Benchmark for Relation Hallucination Evaluation, Analysis and Mitigation in Multimodal Large Language Models
Reefknot: A Comprehensive Benchmark for Relation Hallucination Evaluation, Analysis and Mitigation in Multimodal Large Language ModelsAnnual Meeting of the Association for Computational Linguistics (ACL), 2024
Kening Zheng
Junkai Chen
Yibo Yan
Xin Zou
Xuming Hu
698
17
0
18 Aug 2024
VideoQA in the Era of LLMs: An Empirical Study
VideoQA in the Era of LLMs: An Empirical StudyInternational Journal of Computer Vision (IJCV), 2024
Junbin Xiao
Nanxin Huang
Hangyu Qin
Dongyang Li
Yicong Li
...
Zhulin Tao
Jianxing Yu
Liang Lin
Tat-Seng Chua
Angela Yao
352
24
0
08 Aug 2024
LLaVA-OneVision: Easy Visual Task Transfer
LLaVA-OneVision: Easy Visual Task Transfer
Bo Li
Yuanhan Zhang
Dong Guo
Renrui Zhang
Feng Li
Hao Zhang
Kaichen Zhang
Yanwei Li
Ziwei Liu
Chunyuan Li
MLLMSyDaVLM
570
1,767
0
06 Aug 2024
Alleviating Hallucination in Large Vision-Language Models with Active
  Retrieval Augmentation
Alleviating Hallucination in Large Vision-Language Models with Active Retrieval Augmentation
Cephas Mpungu
Qiyuan Chen
Xiaoye Qu
Jiashuo Sun
G. Mapp
VLMRALMLRM
230
30
0
01 Aug 2024
Paying More Attention to Image: A Training-Free Method for Alleviating
  Hallucination in LVLMs
Paying More Attention to Image: A Training-Free Method for Alleviating Hallucination in LVLMs
Shiping Liu
Kecheng Zheng
Wei Chen
MLLM
254
121
0
31 Jul 2024
Generalized Out-of-Distribution Detection and Beyond in Vision Language Model Era: A Survey
Generalized Out-of-Distribution Detection and Beyond in Vision Language Model Era: A Survey
Atsuyuki Miyai
Jingkang Yang
Jingyang Zhang
Yifei Ming
Sisir Dhakal
...
Yixuan Li
Hai "Helen" Li
Ziwei Liu
Toshihiko Yamasaki
Kiyoharu Aizawa
371
30
0
31 Jul 2024
Interpreting and Mitigating Hallucination in MLLMs through Multi-agent
  Debate
Interpreting and Mitigating Hallucination in MLLMs through Multi-agent Debate
Zheng Lin
Zhenxing Niu
Zhibin Wang
Yinghui Xu
203
15
0
30 Jul 2024
CoMMIT: Coordinated Multimodal Instruction Tuning
CoMMIT: Coordinated Multimodal Instruction Tuning
Junda Wu
Xintong Li
Tong Yu
Yu Wang
Xiang Chen
Jiuxiang Gu
Lina Yao
Julian McAuley
Jingbo Shang
170
4
0
29 Jul 2024
INF-LLaVA: Dual-perspective Perception for High-Resolution Multimodal
  Large Language Model
INF-LLaVA: Dual-perspective Perception for High-Resolution Multimodal Large Language Model
Yiwei Ma
Zhibin Wang
Xiaoshuai Sun
Weihuang Lin
Qiang-feng Zhou
Jiayi Ji
Rongrong Ji
MLLMVLM
241
4
0
23 Jul 2024
The Synergy between Data and Multi-Modal Large Language Models: A Survey
  from Co-Development Perspective
The Synergy between Data and Multi-Modal Large Language Models: A Survey from Co-Development Perspective
Zhen Qin
Daoyuan Chen
Wenhao Zhang
Liuyi Yao
Yilun Huang
Bolin Ding
Yaliang Li
Shuiguang Deng
350
12
0
11 Jul 2024
Merge, Ensemble, and Cooperate! A Survey on Collaborative Strategies in
  the Era of Large Language Models
Merge, Ensemble, and Cooperate! A Survey on Collaborative Strategies in the Era of Large Language Models
Jinliang Lu
Ziliang Pang
Min Xiao
Yaochen Zhu
Rui Xia
Jiajun Zhang
MoMe
393
48
0
08 Jul 2024
Pelican: Correcting Hallucination in Vision-LLMs via Claim Decomposition
  and Program of Thought Verification
Pelican: Correcting Hallucination in Vision-LLMs via Claim Decomposition and Program of Thought Verification
Pritish Sahu
Karan Sikka
Ajay Divakaran
MLLMLRM
220
14
0
02 Jul 2024
Certainly Uncertain: A Benchmark and Metric for Multimodal Epistemic and
  Aleatoric Awareness
Certainly Uncertain: A Benchmark and Metric for Multimodal Epistemic and Aleatoric Awareness
Khyathi Chandu
Linjie Li
Anas Awadalla
Ximing Lu
Jae Sung Park
Jack Hessel
Lijuan Wang
Yejin Choi
322
6
0
02 Jul 2024
Investigating and Mitigating the Multimodal Hallucination Snowballing in
  Large Vision-Language Models
Investigating and Mitigating the Multimodal Hallucination Snowballing in Large Vision-Language Models
Weihong Zhong
Xiaocheng Feng
Liang Zhao
Qiming Li
Lei Huang
Yuxuan Gu
Weitao Ma
Yuan Xu
Bing Qin
MLLM
517
19
0
30 Jun 2024
MLLM as Video Narrator: Mitigating Modality Imbalance in Video Moment
  Retrieval
MLLM as Video Narrator: Mitigating Modality Imbalance in Video Moment Retrieval
Weitong Cai
Jiabo Huang
Shaogang Gong
Hailin Jin
Yang Liu
229
8
0
25 Jun 2024
Evaluating the Quality of Hallucination Benchmarks for Large
  Vision-Language Models
Evaluating the Quality of Hallucination Benchmarks for Large Vision-Language Models
Bei Yan
Jie Zhang
Zheng Yuan
Shiguang Shan
Xilin Chen
VLM
161
13
0
24 Jun 2024
CoMT: Chain-of-Medical-Thought Reduces Hallucination in Medical Report Generation
CoMT: Chain-of-Medical-Thought Reduces Hallucination in Medical Report Generation
Yue Jiang
Jiawei Chen
Dingkang Yang
Mingcheng Li
Shunli Wang
Tong Wu
Ke Li
Lihua Zhang
MLLMLM&MAMedImLRM
152
2
0
17 Jun 2024
SPA-VL: A Comprehensive Safety Preference Alignment Dataset for Vision Language Model
SPA-VL: A Comprehensive Safety Preference Alignment Dataset for Vision Language Model
Yongting Zhang
Lu Chen
Guodong Zheng
Yifeng Gao
Rui Zheng
...
Yu Qiao
Xuanjing Huang
Feng Zhao
Tao Gui
Jing Shao
VLM
518
60
0
17 Jun 2024
Detecting and Evaluating Medical Hallucinations in Large Vision Language
  Models
Detecting and Evaluating Medical Hallucinations in Large Vision Language Models
Jiawei Chen
Dingkang Yang
Tong Wu
Yue Jiang
Xiaolu Hou
Mingcheng Li
Shunli Wang
Dongling Xiao
Ke Li
Lihua Zhang
LM&MAVLM
241
40
0
14 Jun 2024
ClimateIQA: A New Dataset and Benchmark to Advance Vision-Language Models in Meteorology Anomalies Analysis
ClimateIQA: A New Dataset and Benchmark to Advance Vision-Language Models in Meteorology Anomalies Analysis
Jian Chen
Peilin Zhou
Yining Hua
Dading Chong
Meng Cao
Yaowei Li
Wei Chen
Bing Zhu
Junwei Liang
Zixuan Yuan
VLM
458
3
0
14 Jun 2024
Robustness of Structured Data Extraction from In-plane Rotated Documents
  using Multi-Modal Large Language Models (LLM)
Robustness of Structured Data Extraction from In-plane Rotated Documents using Multi-Modal Large Language Models (LLM)
Anjanava Biswas
Wrick Talukdar
157
18
0
13 Jun 2024
CODE: Contrasting Self-generated Description to Combat Hallucination in
  Large Multi-modal Models
CODE: Contrasting Self-generated Description to Combat Hallucination in Large Multi-modal Models
Junho Kim
Hyunjun Kim
Yeonju Kim
Yong Man Ro
MLLM
222
31
0
04 Jun 2024
CheckEmbed: Effective Verification of LLM Solutions to Open-Ended Tasks
CheckEmbed: Effective Verification of LLM Solutions to Open-Ended Tasks
Maciej Besta
Lorenzo Paleari
Marcin Copik
Robert Gerstenberger
Aleš Kubíček
...
Eric Schreiber
Torsten Hoefler
Tomasz Lehmann
H. Niewiadomski
Torsten Hoefler
690
11
0
04 Jun 2024
RITUAL: Random Image Transformations as a Universal Anti-hallucination
  Lever in LVLMs
RITUAL: Random Image Transformations as a Universal Anti-hallucination Lever in LVLMs
Sangmin Woo
Jaehyuk Jang
Donguk Kim
Yubin Choi
Changick Kim
249
0
0
28 May 2024
Alleviating Hallucinations in Large Vision-Language Models through Hallucination-Induced Optimization
Alleviating Hallucinations in Large Vision-Language Models through Hallucination-Induced Optimization
Beitao Chen
Xinyu Lyu
Lianli Gao
Jingkuan Song
Hengtao Shen
MLLM
528
31
0
24 May 2024
Visual Fact Checker: Enabling High-Fidelity Detailed Caption Generation
Visual Fact Checker: Enabling High-Fidelity Detailed Caption Generation
Yunhao Ge
Fangyin Wei
Siddharth Gururani
Nayeon Lee
Xuan Li
Huayu Chen
CoGeDiffM
201
33
0
30 Apr 2024
List Items One by One: A New Data Source and Learning Paradigm for Multimodal LLMs
List Items One by One: A New Data Source and Learning Paradigm for Multimodal LLMs
An Yan
Zhengyuan Yang
Junda Wu
Wanrong Zhu
Jianwei Yang
...
Kevin Qinghong Lin
Jianfeng Wang
Julian McAuley
Jianfeng Gao
Lijuan Wang
LRM
302
25
0
25 Apr 2024
Detecting and Mitigating Hallucination in Large Vision Language Models via Fine-Grained AI Feedback
Detecting and Mitigating Hallucination in Large Vision Language Models via Fine-Grained AI Feedback
Wenyi Xiao
Ziwei Huang
Yaoyao Yu
Wanggui He
Haoyuan Li
Zhelun Yu
Hao Jiang
Leilei Gan
Linchao Zhu
MLLM
357
56
0
22 Apr 2024
FGAIF: Aligning Large Vision-Language Models with Fine-grained AI Feedback
FGAIF: Aligning Large Vision-Language Models with Fine-grained AI Feedback
Liqiang Jing
Xinya Du
392
29
0
07 Apr 2024
Mitigating Hallucinations in Large Vision-Language Models with
  Instruction Contrastive Decoding
Mitigating Hallucinations in Large Vision-Language Models with Instruction Contrastive Decoding
Xintong Wang
Jingheng Pan
Liang Ding
Christian Biemann
MLLM
325
145
0
27 Mar 2024
Multi-Modal Hallucination Control by Visual Information Grounding
Multi-Modal Hallucination Control by Visual Information Grounding
Alessandro Favero
Luca Zancato
Matthew Trager
Siddharth Choudhary
Pramuditha Perera
Alessandro Achille
Ashwin Swaminathan
Stefano Soatto
MLLM
244
148
0
20 Mar 2024
What if...?: Thinking Counterfactual Keywords Helps to Mitigate
  Hallucination in Large Multi-modal Models
What if...?: Thinking Counterfactual Keywords Helps to Mitigate Hallucination in Large Multi-modal Models
Junho Kim
Yeonju Kim
Yonghyun Ro
LRMMLLM
211
9
0
20 Mar 2024
Mitigating Dialogue Hallucination for Large Vision Language Models via
  Adversarial Instruction Tuning
Mitigating Dialogue Hallucination for Large Vision Language Models via Adversarial Instruction Tuning
Dongmin Park
Zhaofang Qian
Guangxing Han
Ser-Nam Lim
MLLM
261
1
0
15 Mar 2024
HALC: Object Hallucination Reduction via Adaptive Focal-Contrast
  Decoding
HALC: Object Hallucination Reduction via Adaptive Focal-Contrast Decoding
Zhaorun Chen
Zhuokai Zhao
Hongyin Luo
Huaxiu Yao
Bo Li
Jiawei Zhou
MLLM
264
132
0
01 Mar 2024
IBD: Alleviating Hallucinations in Large Vision-Language Models via
  Image-Biased Decoding
IBD: Alleviating Hallucinations in Large Vision-Language Models via Image-Biased Decoding
Lanyun Zhu
Deyi Ji
Tianrun Chen
Peng Xu
Jieping Ye
Jun Liu
MLLM
263
90
0
28 Feb 2024
Aligning Modalities in Vision Large Language Models via Preference
  Fine-tuning
Aligning Modalities in Vision Large Language Models via Preference Fine-tuning
Yiyang Zhou
Chenhang Cui
Rafael Rafailov
Chelsea Finn
Huaxiu Yao
VLMMLLM
268
165
0
18 Feb 2024
EFUF: Efficient Fine-grained Unlearning Framework for Mitigating
  Hallucinations in Multimodal Large Language Models
EFUF: Efficient Fine-grained Unlearning Framework for Mitigating Hallucinations in Multimodal Large Language Models
Shangyu Xing
Fei Zhao
Zhen Wu
Tuo An
Weihao Chen
Chunhui Li
Jianbing Zhang
Xinyu Dai
MLLMMU
283
13
0
15 Feb 2024
Mitigating Object Hallucination in Large Vision-Language Models via Image-Grounded Guidance
Mitigating Object Hallucination in Large Vision-Language Models via Image-Grounded Guidance
Linxi Zhao
Yihe Deng
Weitong Zhang
Q. Gu
MLLM
315
30
0
13 Feb 2024
Previous
1234567
Next