ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2311.17911
  4. Cited By
OPERA: Alleviating Hallucination in Multi-Modal Large Language Models
  via Over-Trust Penalty and Retrospection-Allocation
v1v2v3 (latest)

OPERA: Alleviating Hallucination in Multi-Modal Large Language Models via Over-Trust Penalty and Retrospection-Allocation

Computer Vision and Pattern Recognition (CVPR), 2023
29 November 2023
Qidong Huang
Xiao-wen Dong
Pan Zhang
Sijin Yu
Conghui He
Yuan Liu
Dahua Lin
Weiming Zhang
Neng H. Yu
    MLLM
ArXiv (abs)PDFHTMLHuggingFace (2 upvotes)Github (341★)

Papers citing "OPERA: Alleviating Hallucination in Multi-Modal Large Language Models via Over-Trust Penalty and Retrospection-Allocation"

50 / 238 papers shown
Mitigating Object and Action Hallucinations in Multimodal LLMs via Self-Augmented Contrastive Alignment
Mitigating Object and Action Hallucinations in Multimodal LLMs via Self-Augmented Contrastive Alignment
Kai-Po Chang
Wei-Yuan Cheng
Chi-Pin Huang
Fu-En Yang
Yu-Jie Wang
256
1
0
04 Dec 2025
SEASON: Mitigating Temporal Hallucination in Video Large Language Models via Self-Diagnostic Contrastive Decoding
SEASON: Mitigating Temporal Hallucination in Video Large Language Models via Self-Diagnostic Contrastive Decoding
Chang-Hsun Wu
Kai-Po Chang
Yu-Yang Sheng
Hung-Kai Chung
Kuei-Chun Wang
Yu-Jie Wang
MLLM
222
0
0
04 Dec 2025
V-ITI: Mitigating Hallucinations in Multimodal Large Language Models via Visual Inference-Time Intervention
V-ITI: Mitigating Hallucinations in Multimodal Large Language Models via Visual Inference-Time Intervention
Nan Sun
Zhenyu Zhang
Xixun Lin
Kun Wang
Yanmin Shang
...
Shuohuan Wang
Yu Sun
H. Wu
Haifeng Wang
Yanan Cao
MLLMVLM
132
0
0
03 Dec 2025
InEx: Hallucination Mitigation via Introspection and Cross-Modal Multi-Agent Collaboration
InEx: Hallucination Mitigation via Introspection and Cross-Modal Multi-Agent Collaboration
Zhongyu Yang
Yingfang Yuan
Xuanming Jiang
Baoyi An
Wei Pang
LLMAGHILMLRM
144
0
0
02 Dec 2025
Med-VCD: Mitigating Hallucination for Medical Large Vision Language Models through Visual Contrastive Decoding
Med-VCD: Mitigating Hallucination for Medical Large Vision Language Models through Visual Contrastive DecodingComputers in Biology and Medicine (Comput. Biol. Med.), 2025
Zahra Mahdavi
Zahra Khodakaramimaghsoud
Hooman Khaloo
Sina Bakhshandeh Taleshani
Erfan Hashemi
Javad Mirzapour Kaleybar
Omid Nejati Manzari
MLLMVLM
225
0
0
01 Dec 2025
Tell Model Where to Look: Mitigating Hallucinations in MLLMs by Vision-Guided Attention
Tell Model Where to Look: Mitigating Hallucinations in MLLMs by Vision-Guided Attention
Jianfei Zhao
Feng Zhang
Xin Sun
Chong Feng
Zhixing Tan
MLLMVLM
185
0
0
25 Nov 2025
Trust in Vision-Language Models: Insights from a Participatory User Workshop
Trust in Vision-Language Models: Insights from a Participatory User Workshop
Agnese Chiatti
Lara Shibelski Godoy Piccolo
Sara Bernardini
Matteo Matteucci
Viola Schiaffonati
57
0
0
17 Nov 2025
VOPE: Revisiting Hallucination of Vision-Language Models in Voluntary Imagination Task
VOPE: Revisiting Hallucination of Vision-Language Models in Voluntary Imagination Task
Xingming Long
Jie M. Zhang
Shiguang Shan
Xilin Chen
MLLMLRM
149
0
0
17 Nov 2025
What Color Is It? A Text-Interference Multimodal Hallucination Benchmark
What Color Is It? A Text-Interference Multimodal Hallucination Benchmark
Jinkun Zhao
Lei Huang
Haixin Ge
Wenjun Wu
VLM
237
1
0
17 Nov 2025
Suppressing VLM Hallucinations with Spectral Representation Filtering
Suppressing VLM Hallucinations with Spectral Representation Filtering
Ameen Ali
Tamim Zoabi
Lior Wolf
142
0
0
15 Nov 2025
Causal Tracing of Object Representations in Large Vision Language Models: Mechanistic Interpretability and Hallucination Mitigation
Causal Tracing of Object Representations in Large Vision Language Models: Mechanistic Interpretability and Hallucination Mitigation
Qiming Li
Zekai Ye
Xiaocheng Feng
Weihong Zhong
Weitao Ma
Xiachong Feng
200
4
0
08 Nov 2025
SteerVLM: Robust Model Control through Lightweight Activation Steering for Vision Language Models
SteerVLM: Robust Model Control through Lightweight Activation Steering for Vision Language Models
Anushka Sivakumar
Andrew Zhang
Zaber Ibn Abdul Hakim
Chris Thomas
LLMSV
204
0
0
30 Oct 2025
Towards Fine-Grained Human Motion Video Captioning
Towards Fine-Grained Human Motion Video Captioning
Guorui Song
Guocun Wang
Zhe Huang
Jing Lin
Xuefei Zhe
Jian Li
Haoqian Wang
88
0
0
24 Oct 2025
Gaze-VLM:Bridging Gaze and VLMs through Attention Regularization for Egocentric Understanding
Gaze-VLM:Bridging Gaze and VLMs through Attention Regularization for Egocentric Understanding
Anupam Pani
Yanchao Yang
107
0
0
24 Oct 2025
VL-SAE: Interpreting and Enhancing Vision-Language Alignment with a Unified Concept Set
VL-SAE: Interpreting and Enhancing Vision-Language Alignment with a Unified Concept Set
Shufan Shen
Junshu Sun
Qingming Huang
Shuhui Wang
142
1
0
24 Oct 2025
Why LVLMs Are More Prone to Hallucinations in Longer Responses: The Role of Context
Why LVLMs Are More Prone to Hallucinations in Longer Responses: The Role of Context
Ge Zheng
Jiaye Qian
Jiajin Tang
Sibei Yang
94
3
0
23 Oct 2025
PruneHal: Reducing Hallucinations in Multi-modal Large Language Models through Adaptive KV Cache Pruning
PruneHal: Reducing Hallucinations in Multi-modal Large Language Models through Adaptive KV Cache Pruning
Fengyuan Sun
Hui Chen
Xinhao Xu
Dandan Zheng
Jingdong Chen
Jun Zhou
Jungong Han
Guiguang Ding
VLM
120
0
0
22 Oct 2025
Med-VRAgent: A Framework for Medical Visual Reasoning-Enhanced Agents
Med-VRAgent: A Framework for Medical Visual Reasoning-Enhanced Agents
Guangfu Guo
Xiaoqian Lu
Yue Feng
LRM
180
0
0
21 Oct 2025
Beyond Single Models: Mitigating Multimodal Hallucinations via Adaptive Token Ensemble Decoding
Beyond Single Models: Mitigating Multimodal Hallucinations via Adaptive Token Ensemble Decoding
Jinlin Li
Y. X. R. Wang
Yifei Yuan
Xiao Zhou
Y. Zhang
Xixian Yong
Yefeng Zheng
X. Wu
MLLM
151
0
0
21 Oct 2025
SHIELD: Suppressing Hallucinations In LVLM Encoders via Bias and Vulnerability Defense
SHIELD: Suppressing Hallucinations In LVLM Encoders via Bias and Vulnerability Defense
Y. Huang
Liang Shi
Yitian Zhang
Yi Tian Xu
Yun Fu
AAML
92
0
0
18 Oct 2025
Watermarking for Factuality: Guiding Vision-Language Models Toward Truth via Tri-layer Contrastive Decoding
Watermarking for Factuality: Guiding Vision-Language Models Toward Truth via Tri-layer Contrastive Decoding
Kyungryul Back
Seongbeom Park
Milim Kim
Mincheol Kwon
SangHyeok Lee
Hyunyoung Lee
Junhee Cho
Seunghyun Park
Jinkyu Kim
115
0
0
16 Oct 2025
Self-Augmented Visual Contrastive Decoding
Self-Augmented Visual Contrastive Decoding
Eun Woo Im
M. K. Ali
Vivek Gupta
133
0
0
15 Oct 2025
The Mechanistic Emergence of Symbol Grounding in Language Models
The Mechanistic Emergence of Symbol Grounding in Language Models
Shuyu Wu
Ziqiao Ma
Xiaoxi Luo
Yidong Huang
Josue Torres-Fonseca
Freda Shi
Joyce Chai
LRM
180
2
0
15 Oct 2025
DSAS: A Universal Plug-and-Play Framework for Attention Optimization in Multi-Document Question Answering
DSAS: A Universal Plug-and-Play Framework for Attention Optimization in Multi-Document Question Answering
Jiakai Li
Rongzheng Wang
Yizhuo Ma
Shuang Liang
Guangchun Luo
Ke Qin
129
0
0
14 Oct 2025
Mitigating Hallucination in Multimodal Reasoning via Functional Attention Control
Mitigating Hallucination in Multimodal Reasoning via Functional Attention Control
H. Lu
Bolun Chu
Weiye Fu
Guoshun Nan
Junning Liu
Minghui Pan
Qiankun Li
Yi Yu
Hua Wang
Kun Wang
LRM
129
0
0
11 Oct 2025
Towards Better & Faster Autoregressive Image Generation: From the Perspective of Entropy
Towards Better & Faster Autoregressive Image Generation: From the Perspective of Entropy
Xiaoxiao Ma
Feng Zhao
Pengyang Ling
Haibo Qiu
Zhixiang Wei
Hu Yu
Jie Huang
Zhixiong Zeng
Lin Ma
155
2
0
10 Oct 2025
To Sink or Not to Sink: Visual Information Pathways in Large Vision-Language Models
To Sink or Not to Sink: Visual Information Pathways in Large Vision-Language Models
Jiayun Luo
Wan-Cyuan Fan
Lyuyang Wang
Xiangteng He
Tanzila Rahman
Purang Abolmaesumi
Leonid Sigal
LRM
140
0
0
09 Oct 2025
Zero-Shot Fine-Grained Image Classification Using Large Vision-Language Models
Zero-Shot Fine-Grained Image Classification Using Large Vision-Language Models
Md. Atabuzzaman
Andrew Zhang
Chris Thomas
MLLMVLM
140
0
0
04 Oct 2025
Don't Just Chase "Highlighted Tokens" in MLLMs: Revisiting Visual Holistic Context Retention
Don't Just Chase "Highlighted Tokens" in MLLMs: Revisiting Visual Holistic Context Retention
Xin Zou
Di Lu
Yizhou Wang
Yibo Yan
Yuanhuiyi Lyu
Xu Zheng
Linfeng Zhang
Xuming Hu
VLM
281
5
0
03 Oct 2025
MaskCD: Mitigating LVLM Hallucinations by Image Head Masked Contrastive Decoding
MaskCD: Mitigating LVLM Hallucinations by Image Head Masked Contrastive Decoding
Jingyuan Deng
Yujiu Yang
MLLM
178
1
0
03 Oct 2025
Understanding Sensitivity of Differential Attention through the Lens of Adversarial Robustness
Understanding Sensitivity of Differential Attention through the Lens of Adversarial Robustness
Tsubasa Takahashi
Shojiro Yamabe
Futa Waseda
Kento Sasaki
AAML
152
0
0
01 Oct 2025
From Perception to Cognition: A Survey of Vision-Language Interactive Reasoning in Multimodal Large Language Models
From Perception to Cognition: A Survey of Vision-Language Interactive Reasoning in Multimodal Large Language Models
Chenyue Zhou
Mingxuan Wang
Yanbiao Ma
Chenxu Wu
Wanyi Chen
...
Guoli Jia
Lingling Li
Z. Lu
Y. Lu
Wenhan Luo
LRM
447
9
0
29 Sep 2025
Mitigating Hallucination in Multimodal LLMs with Layer Contrastive Decoding
Mitigating Hallucination in Multimodal LLMs with Layer Contrastive Decoding
Bingkui Tong
Jiaer Xia
Kaiyang Zhou
MLLM
176
1
0
29 Sep 2025
Latent Visual Reasoning
Latent Visual Reasoning
Bangzheng Li
Ximeng Sun
Jiang-Long Liu
Ze Wang
Jialian Wu
Xiaodong Yu
Hao Chen
Emad Barsoum
Muhao Chen
Zicheng Liu
LRMVLM
200
6
0
29 Sep 2025
Uncovering Grounding IDs: How External Cues Shape Multimodal Binding
Uncovering Grounding IDs: How External Cues Shape Multimodal Binding
Hosein Hasani
Amirmohammad Izadi
Fatemeh Askari
Mobin Bagherian
Sadegh Mohammadian
Mohammad Izadi
M. Baghshah
317
0
0
28 Sep 2025
Self-Consistency as a Free Lunch: Reducing Hallucinations in Vision-Language Models via Self-Reflection
Self-Consistency as a Free Lunch: Reducing Hallucinations in Vision-Language Models via Self-Reflection
Mingfei Han
Haihong Hao
Jinxing Zhou
Zhihui Li
Yuhui Zheng
XueQing Deng
Linjie Yang
Xiaojun Chang
HILMVLM
116
0
0
27 Sep 2025
MMPB: It's Time for Multi-Modal Personalization
MMPB: It's Time for Multi-Modal Personalization
Jaeik Kim
Woojin Kim
Woohyeon Park
Jaeyoung Do
VLM
190
0
0
26 Sep 2025
LLM-based Agents Suffer from Hallucinations: A Survey of Taxonomy, Methods, and Directions
LLM-based Agents Suffer from Hallucinations: A Survey of Taxonomy, Methods, and Directions
Xixun Lin
Yucheng Ning
Jingwen Zhang
Yan Dong
Y. Liu
...
Bin Wang
Yanan Cao
Kai-xiang Chen
Songlin Hu
Li Guo
LLMAGLRM
335
4
0
23 Sep 2025
Diving into Mitigating Hallucinations from a Vision Perspective for Large Vision-Language Models
Diving into Mitigating Hallucinations from a Vision Perspective for Large Vision-Language Models
Weihang Wang
Xinhao Li
Ziyue Wang
Yan Pang
Jielei Zhang
Peiyi Li
Qiang Zhang
Longwen Gao
VLM
168
1
0
17 Sep 2025
Cross-Layer Vision Smoothing: Enhancing Visual Understanding via Sustained Focus on Key Objects in Large Vision-Language Models
Cross-Layer Vision Smoothing: Enhancing Visual Understanding via Sustained Focus on Key Objects in Large Vision-Language Models
Jianfei Zhao
Feng Zhang
Xin Sun
Lingxing Kong
Zhixing Tan
137
1
0
16 Sep 2025
Mitigating Hallucinations in Large Vision-Language Models by Self-Injecting Hallucinations
Mitigating Hallucinations in Large Vision-Language Models by Self-Injecting Hallucinations
Y. Lu
Ziqi Zhang
Chunfeng Yuan
Jun Gao
Congxuan Zhang
Xiaojuan Qi
Bing Li
Weiming Hu
MLLMVLM
97
0
0
14 Sep 2025
Measuring Epistemic Humility in Multimodal Large Language Models
Measuring Epistemic Humility in Multimodal Large Language Models
Bingkui Tong
Jiaer Xia
Sifeng Shang
Kaiyang Zhou
HILM
136
2
0
11 Sep 2025
Tracing and Mitigating Hallucinations in Multimodal LLMs via Dynamic Attention Localization
Tracing and Mitigating Hallucinations in Multimodal LLMs via Dynamic Attention Localization
Tiancheng Yang
L. Zhang
J. Lin
Guimin Hu
Haiyan Zhao
Lijie Hu
168
0
0
09 Sep 2025
Cross-Layer Attention Probing for Fine-Grained Hallucination Detection
Cross-Layer Attention Probing for Fine-Grained Hallucination Detection
Malavika Suresh
Rahaf Aljundi
Ikechukwu Nkisi-Orji
Nirmalie Wiratunga
HILM
170
1
0
04 Sep 2025
Mitigating Multimodal Hallucinations via Gradient-based Self-Reflection
Mitigating Multimodal Hallucinations via Gradient-based Self-Reflection
Shan Wang
Maying Shen
Nadine Chang
Chuong H. Nguyen
Hongdong Li
J. Álvarez
259
0
0
03 Sep 2025
Unveiling the Response of Large Vision-Language Models to Visually Absent Tokens
Unveiling the Response of Large Vision-Language Models to Visually Absent Tokens
Sohee Kim
Soohyun Ryu
Joonhyung Park
Eunho Yang
151
0
0
03 Sep 2025
OmniDPO: A Preference Optimization Framework to Address Omni-Modal Hallucination
OmniDPO: A Preference Optimization Framework to Address Omni-Modal Hallucination
Junzhe Chen
Tianshu Zhang
Shiyu Huang
Yuwei Niu
Chao Sun
Rongzhou Zhang
G. Zhou
Lijie Wen
Xuming Hu
MLLM
185
0
0
31 Aug 2025
Negative Matters: Multi-Granularity Hard-Negative Synthesis and Anchor-Token-Aware Pooling for Enhanced Text Embeddings
Negative Matters: Multi-Granularity Hard-Negative Synthesis and Anchor-Token-Aware Pooling for Enhanced Text EmbeddingsAnnual Meeting of the Association for Computational Linguistics (ACL), 2025
Tengyu Pan
Zhichao Duan
Zhenyu Li
Bowen Dong
Ning Liu
Xiuxing Li
Jianyong Wang
104
1
0
31 Aug 2025
Improving Alignment in LVLMs with Debiased Self-Judgment
Improving Alignment in LVLMs with Debiased Self-Judgment
Sihan Yang
Chenhang Cui
Zihao Zhao
Yiyang Zhou
Weilong Yan
Ying Wei
Huaxiu Yao
209
0
0
28 Aug 2025
Mitigating Hallucinations in Multimodal LLMs via Object-aware Preference Optimization
Mitigating Hallucinations in Multimodal LLMs via Object-aware Preference Optimization
Alberto Compagnoni
Davide Caffagni
Nicholas Moratelli
Lorenzo Baraldi
Marcella Cornia
Rita Cucchiara
MLLM
197
1
0
27 Aug 2025
12345
Next