ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2407.21771
  4. Cited By
Paying More Attention to Image: A Training-Free Method for Alleviating
  Hallucination in LVLMs

Paying More Attention to Image: A Training-Free Method for Alleviating Hallucination in LVLMs

31 July 2024
Shiping Liu
Kecheng Zheng
Wei Chen
    MLLM
ArXiv (abs)PDFHTMLGithub

Papers citing "Paying More Attention to Image: A Training-Free Method for Alleviating Hallucination in LVLMs"

50 / 71 papers shown
Through the Magnifying Glass: Adaptive Perception Magnification for Hallucination-Free VLM Decoding
Through the Magnifying Glass: Adaptive Perception Magnification for Hallucination-Free VLM Decoding
Shunqi Mao
Chaoyi Zhang
Weidong Cai
MLLM
1.2K
6
0
10 Apr 2026
V-ITI: Mitigating Hallucinations in Multimodal Large Language Models via Visual Inference-Time Intervention
V-ITI: Mitigating Hallucinations in Multimodal Large Language Models via Visual Inference-Time Intervention
Nan Sun
Zhenyu Zhang
Xixun Lin
Kun Wang
Yanmin Shang
...
Shuohuan Wang
Yu Sun
H. Wu
Haifeng Wang
Yanan Cao
MLLMVLM
187
1
0
03 Dec 2025
Med-VCD: Mitigating Hallucination for Medical Large Vision Language Models through Visual Contrastive Decoding
Med-VCD: Mitigating Hallucination for Medical Large Vision Language Models through Visual Contrastive DecodingComputers in Biology and Medicine (Comput. Biol. Med.), 2025
Zahra Mahdavi
Zahra Khodakaramimaghsoud
Hooman Khaloo
Sina Bakhshandeh Taleshani
Erfan Hashemi
Javad Mirzapour Kaleybar
Omid Nejati Manzari
MLLMVLM
305
1
0
01 Dec 2025
Tell Model Where to Look: Mitigating Hallucinations in MLLMs by Vision-Guided Attention
Tell Model Where to Look: Mitigating Hallucinations in MLLMs by Vision-Guided Attention
Jianfei Zhao
Feng Zhang
Xin Sun
Chong Feng
Zhixing Tan
MLLMLRM
253
1
0
25 Nov 2025
Intervene-All-Paths: Unified Mitigation of LVLM Hallucinations across Alignment Formats
Intervene-All-Paths: Unified Mitigation of LVLM Hallucinations across Alignment Formats
Jiaye Qian
Ge Zheng
Yuchen Zhu
Sibei Yang
MLLM
402
3
0
21 Nov 2025
Adaptive Residual-Update Steering for Low-Overhead Hallucination Mitigation in Large Vision Language Models
Adaptive Residual-Update Steering for Low-Overhead Hallucination Mitigation in Large Vision Language Models
Zhengtao Zou
Ya Gao
Jiarui Guan
Bin Li
Pekka Marttinen
172
1
0
13 Nov 2025
Capturing Gaze Shifts for Guidance: Cross-Modal Fusion Enhancement for VLM Hallucination Mitigation
Capturing Gaze Shifts for Guidance: Cross-Modal Fusion Enhancement for VLM Hallucination Mitigation
Zheng Qi
Chao Shang
Evangelia Spiliopoulou
Nikolaos Pappas
220
4
0
24 Oct 2025
Why LVLMs Are More Prone to Hallucinations in Longer Responses: The Role of Context
Why LVLMs Are More Prone to Hallucinations in Longer Responses: The Role of Context
Ge Zheng
Jiaye Qian
Jiajin Tang
Sibei Yang
145
9
0
23 Oct 2025
PruneHal: Reducing Hallucinations in Multi-modal Large Language Models through Adaptive KV Cache Pruning
PruneHal: Reducing Hallucinations in Multi-modal Large Language Models through Adaptive KV Cache Pruning
Fengyuan Sun
Hui Chen
Xinhao Xu
Dandan Zheng
Jingdong Chen
Jun Zhou
Jungong Han
Guiguang Ding
VLM
146
1
0
22 Oct 2025
Seeing but Not Believing: Probing the Disconnect Between Visual Attention and Answer Correctness in VLMs
Seeing but Not Believing: Probing the Disconnect Between Visual Attention and Answer Correctness in VLMs
Zhining Liu
Ziyi Chen
Hui Liu
Chen Luo
Xianfeng Tang
...
Zhenwei Dai
Zhan Shi
Tianxin Wei
Benoit Dumoulin
Hanghang Tong
LRM
165
20
0
20 Oct 2025
Watermarking for Factuality: Guiding Vision-Language Models Toward Truth via Tri-layer Contrastive Decoding
Watermarking for Factuality: Guiding Vision-Language Models Toward Truth via Tri-layer Contrastive Decoding
Kyungryul Back
Seongbeom Park
Milim Kim
Mincheol Kwon
SangHyeok Lee
Hyunyoung Lee
Junhee Cho
Seunghyun Park
Jinkyu Kim
223
0
0
16 Oct 2025
Zero-Shot Fine-Grained Image Classification Using Large Vision-Language Models
Zero-Shot Fine-Grained Image Classification Using Large Vision-Language Models
Md. Atabuzzaman
Andrew Zhang
Chris Thomas
MLLMVLM
189
1
0
04 Oct 2025
MaskCD: Mitigating LVLM Hallucinations by Image Head Masked Contrastive Decoding
MaskCD: Mitigating LVLM Hallucinations by Image Head Masked Contrastive Decoding
Jingyuan Deng
Yujiu Yang
MLLM
213
1
0
03 Oct 2025
CCD: Mitigating Hallucinations in Radiology MLLMs via Clinical Contrastive Decoding
CCD: Mitigating Hallucinations in Radiology MLLMs via Clinical Contrastive Decoding
Xi Zhang
Zaiqiao Meng
Jake Lever
Edmond S. L. Ho
MedIm
313
1
0
27 Sep 2025
Understanding Language Prior of LVLMs by Contrasting Chain-of-Embedding
Understanding Language Prior of LVLMs by Contrasting Chain-of-Embedding
Lin Long
Changdae Oh
Seongheon Park
Yixuan Li
VLMMLLM
239
3
1
27 Sep 2025
Visual Multi-Agent System: Mitigating Hallucination Snowballing via Visual Flow
Visual Multi-Agent System: Mitigating Hallucination Snowballing via Visual Flow
Xinlei Yu
C. Xu
Guibin Zhang
Yongbo He
Zhangquan Chen
...
Jiangning Zhang
Yue Liao
Xiaobin Hu
Yu-Gang Jiang
Shuicheng Yan
319
11
0
26 Sep 2025
Pay More Attention To Audio: Mitigating Imbalance of Cross-Modal Attention in Large Audio Language Models
Pay More Attention To Audio: Mitigating Imbalance of Cross-Modal Attention in Large Audio Language Models
Junyu Wang
Ziyang Ma
Zhengding Luo
Tianrui Wang
Meng Ge
Xiaobao Wang
Longbiao Wang
AuLLM
137
1
0
23 Sep 2025
Cross-Layer Vision Smoothing: Enhancing Visual Understanding via Sustained Focus on Key Objects in Large Vision-Language Models
Cross-Layer Vision Smoothing: Enhancing Visual Understanding via Sustained Focus on Key Objects in Large Vision-Language Models
Jianfei Zhao
Feng Zhang
Xin Sun
Lingxing Kong
Zhixing Tan
202
1
0
16 Sep 2025
ChartGaze: Enhancing Chart Understanding in LVLMs with Eye-Tracking Guided Attention Refinement
ChartGaze: Enhancing Chart Understanding in LVLMs with Eye-Tracking Guided Attention Refinement
Ali Salamatian
Amirhossein Abaskohi
Wan-Cyuan Fan
Mir Rayat Imtiaz Hossain
Leonid Sigal
Giuseppe Carenini
168
3
0
16 Sep 2025
Tracing and Mitigating Hallucinations in Multimodal LLMs via Dynamic Attention Localization
Tracing and Mitigating Hallucinations in Multimodal LLMs via Dynamic Attention Localization
Tiancheng Yang
L. Zhang
J. Lin
Guimin Hu
Haiyan Zhao
Lijie Hu
357
0
0
09 Sep 2025
SPECS: Specificity-Enhanced CLIP-Score for Long Image Caption Evaluation
SPECS: Specificity-Enhanced CLIP-Score for Long Image Caption Evaluation
Xiaofu Chen
Israfel Salazar
Yova Kementchedjhieva
347
2
0
04 Sep 2025
Unveiling the Response of Large Vision-Language Models to Visually Absent Tokens
Unveiling the Response of Large Vision-Language Models to Visually Absent Tokens
Sohee Kim
Soohyun Ryu
Joonhyung Park
Eunho Yang
179
0
0
03 Sep 2025
Mitigating Multimodal Hallucinations via Gradient-based Self-Reflection
Mitigating Multimodal Hallucinations via Gradient-based Self-Reflection
Shan Wang
Maying Shen
Nadine Chang
Chuong H. Nguyen
Hongdong Li
J. Álvarez
359
0
0
03 Sep 2025
OmniDPO: A Preference Optimization Framework to Address Omni-Modal Hallucination
OmniDPO: A Preference Optimization Framework to Address Omni-Modal Hallucination
Junzhe Chen
Tianshu Zhang
Shiyu Huang
Yuwei Niu
Chao Sun
Rongzhou Zhang
G. Zhou
Lijie Wen
Xuming Hu
MLLM
228
4
0
31 Aug 2025
GLSim: Detecting Object Hallucinations in LVLMs via Global-Local Similarity
GLSim: Detecting Object Hallucinations in LVLMs via Global-Local Similarity
Seongheon Park
Yixuan Li
225
3
0
27 Aug 2025
Benchmarking and Bridging Emotion Conflicts for Multimodal Emotion Reasoning
Benchmarking and Bridging Emotion Conflicts for Multimodal Emotion Reasoning
Zhiyuan Han
Beier Zhu
Yanlong Xu
Peipei Song
Xun Yang
240
12
0
02 Aug 2025
MIHBench: Benchmarking and Mitigating Multi-Image Hallucinations in Multimodal Large Language Models
MIHBench: Benchmarking and Mitigating Multi-Image Hallucinations in Multimodal Large Language Models
Jiale Li
Mingrui Wu
Zixiang Jin
Hao Chen
Jinfa Huang
Xiaoshuai Sun
Liujuan Cao
Rongrong Ji
VLM
310
3
0
01 Aug 2025
TARS: MinMax Token-Adaptive Preference Strategy for Hallucination Reduction in MLLMs
TARS: MinMax Token-Adaptive Preference Strategy for Hallucination Reduction in MLLMs
Kejia Zhang
Keda Tao
Zhiming Luo
Chang Liu
Jiasheng Tang
Huan Wang
379
0
0
29 Jul 2025
OW-CLIP: Data-Efficient Visual Supervision for Open-World Object Detection via Human-AI Collaboration
OW-CLIP: Data-Efficient Visual Supervision for Open-World Object Detection via Human-AI Collaboration
Junwen Duan
Wei Xue
Ziyao Kang
Shixia Liu
Jiazhi Xia
VLM
219
0
0
26 Jul 2025
MCA-LLaVA: Manhattan Causal Attention for Reducing Hallucination in Large Vision-Language Models
MCA-LLaVA: Manhattan Causal Attention for Reducing Hallucination in Large Vision-Language Models
Qiyan Zhao
Xiaofeng Zhang
Yiheng Li
Yun Xing
Xiaosong Yuan
Feilong Tang
Sinan Fan
Xuhang Chen
Xuyao Zhang
Dahan Wang
407
7
0
12 Jul 2025
Not All Attention Heads Are What You Need: Refining CLIP's Image Representation with Attention Ablation
Not All Attention Heads Are What You Need: Refining CLIP's Image Representation with Attention Ablation
Feng Lin
Marco Chen
Haokui Zhang
Xiaotian Yu
Guangming Lu
Rong Xiao
158
1
0
01 Jul 2025
ASCD: Attention-Steerable Contrastive Decoding for Reducing Hallucination in MLLM
ASCD: Attention-Steerable Contrastive Decoding for Reducing Hallucination in MLLM
Yujun Wang
Aniri
Jinhe Bi
Soeren Pirk
Yunpu Ma
MLLM
410
20
0
17 Jun 2025
Revisit What You See: Disclose Language Prior in Vision Tokens for LVLM Decoding
Revisit What You See: Disclose Language Prior in Vision Tokens for LVLM Decoding
Beomsik Cho
Jaehyung Kim
350
0
0
11 Jun 2025
CoMemo: LVLMs Need Image Context with Image Memory
CoMemo: LVLMs Need Image Context with Image Memory
Shi-Qi Liu
Weijie Su
Xizhou Zhu
Wenhai Wang
Jifeng Dai
VLM
276
3
0
06 Jun 2025
LLMs Can Compensate for Deficiencies in Visual Representations
LLMs Can Compensate for Deficiencies in Visual Representations
Sho Takishita
Jay Gala
Abdelrahman Mohamed
Kentaro Inui
Yova Kementchedjhieva
VLM
259
1
0
05 Jun 2025
Mitigating Hallucinations in Large Vision-Language Models via Entity-Centric Multimodal Preference Optimization
Mitigating Hallucinations in Large Vision-Language Models via Entity-Centric Multimodal Preference Optimization
Jiulong Wu
Zhengliang Shi
Shuaiqiang Wang
J. Huang
Dawei Yin
Lingyong Yan
Min Cao
Min Zhang
MLLM
397
8
0
04 Jun 2025
CLAIM: Mitigating Multilingual Object Hallucination in Large Vision-Language Models with Cross-Lingual Attention Intervention
CLAIM: Mitigating Multilingual Object Hallucination in Large Vision-Language Models with Cross-Lingual Attention InterventionAnnual Meeting of the Association for Computational Linguistics (ACL), 2025
Zekai Ye
Qiming Li
Xiaocheng Feng
L. Qin
Yichong Huang
...
Zhirui Zhang
Yunfei Lu
Duyu Tang
Dandan Tu
Bing Qin
VLMLRM
210
12
0
03 Jun 2025
BIMA: Bijective Maximum Likelihood Learning Approach to Hallucination Prediction and Mitigation in Large Vision-Language Models
BIMA: Bijective Maximum Likelihood Learning Approach to Hallucination Prediction and Mitigation in Large Vision-Language Models
Huu-Thien Tran
Thanh-Dat Truong
Khoa Luu
MLLM
218
1
0
30 May 2025
Qwen Look Again: Guiding Vision-Language Reasoning Models to Re-attention Visual Information
Qwen Look Again: Guiding Vision-Language Reasoning Models to Re-attention Visual Information
Xu Chu
Xinrong Chen
Guanyu Wang
Zhijie Tan
Kui Huang
Wenyu Lv
Tong Mo
Weiping Li
LRMVLM
437
16
0
29 May 2025
Mitigating Hallucination in Large Vision-Language Models via Adaptive Attention Calibration
Mitigating Hallucination in Large Vision-Language Models via Adaptive Attention Calibration
Mehrdad Fazli
Bowen Wei
Ahmet Sari
Ziwei Zhu
VLM
557
5
0
27 May 2025
MLLMs are Deeply Affected by Modality Bias
MLLMs are Deeply Affected by Modality Bias
Xu Zheng
Chenfei Liao
Yuqian Fu
Kaiyu Lei
Yuanhuiyi Lyu
...
Yu Jiang
Andrii Zadaianchuk
Dacheng Tao
Luc Van Gool
Xuming Hu
395
24
0
24 May 2025
Seeing Far and Clearly: Mitigating Hallucinations in MLLMs with Attention Causal Decoding
Seeing Far and Clearly: Mitigating Hallucinations in MLLMs with Attention Causal DecodingComputer Vision and Pattern Recognition (CVPR), 2025
Feilong Tang
Chengzhi Liu
Zhongxing Xu
Ming Hu
Zelin Peng
...
Minquan Lin
Yifan Peng
Xuelian Cheng
Imran Razzak
Zongyuan Ge
425
35
0
22 May 2025
Mitigating Hallucinations in Vision-Language Models through Image-Guided Head Suppression
Mitigating Hallucinations in Vision-Language Models through Image-Guided Head Suppression
Sreetama Sarkar
Yue Che
Alex Gavin
Peter A. Beerel
Souvik Kundu
MLLMVLM
354
10
0
22 May 2025
Make LVLMs Focus: Context-Aware Attention Modulation for Better Multimodal In-Context Learning
Make LVLMs Focus: Context-Aware Attention Modulation for Better Multimodal In-Context Learning
Yanshu Li
JianJiang Yang
Ziteng Yang
Bozheng Li
Yi Cao
...
Ligong Han
Yingjie Victor Chen
Songlin Fei
Dongfang Liu
Ruixiang Tang
487
8
0
21 May 2025
How Do Large Vision-Language Models See Text in Image? Unveiling the Distinctive Role of OCR Heads
How Do Large Vision-Language Models See Text in Image? Unveiling the Distinctive Role of OCR Heads
Ingeol Baek
Hwan Chang
Sunghyun Ryu
Hwanhee Lee
251
5
0
21 May 2025
ZeroTuning: Unlocking the Initial Token's Power to Enhance Large Language Models Without Training
ZeroTuning: Unlocking the Initial Token's Power to Enhance Large Language Models Without Training
Feijiang Han
Xiaodong Yu
Jianheng Tang
Delip Rao
Weihua Du
Lyle Ungar
522
11
0
16 May 2025
Decoupling Contrastive Decoding: Robust Hallucination Mitigation in Multimodal Large Language Models
Decoupling Contrastive Decoding: Robust Hallucination Mitigation in Multimodal Large Language Models
Wei Chen
Xin Yan
Bin Wen
Fan Yang
Yan Li
Di Zhang
Long Chen
MLLM
515
2
0
09 Apr 2025
Exploring Hallucination of Large Multimodal Models in Video Understanding: Benchmark, Analysis and Mitigation
Exploring Hallucination of Large Multimodal Models in Video Understanding: Benchmark, Analysis and Mitigation
Hongcheng Gao
Jiashu Qu
Jingyi Tang
Baolong Bi
Yi Liu
Hongyu Chen
Li Liang
Li Su
Qingming Huang
MLLMVLMLRM
534
20
0
25 Mar 2025
Mitigating Object Hallucinations in MLLMs via Multi-Frequency Perturbations
Mitigating Object Hallucinations in MLLMs via Multi-Frequency Perturbations
Shuo Li
Jiajun Sun
Guodong Zheng
Xiaoran Fan
Yujiong Shen
...
Wenming Tan
Changzhi Sun
Tao Gui
Tao Gui
Qi Zhang
AAMLVLM
426
9
0
19 Mar 2025
Grounded Chain-of-Thought for Multimodal Large Language Models
Grounded Chain-of-Thought for Multimodal Large Language Models
Qiong Wu
Xiangcong Yang
Weihao Ye
Chenxin Fang
Baiyang Song
Xiaoshuai Sun
Rongrong Ji
LRM
592
32
0
17 Mar 2025
12
Next
Page 1 of 2