ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2406.04236
  4. Cited By
Understanding Information Storage and Transfer in Multi-modal Large
  Language Models

Understanding Information Storage and Transfer in Multi-modal Large Language Models

Neural Information Processing Systems (NeurIPS), 2024
6 June 2024
Samyadeep Basu
Martin Grayson
C. Morrison
Besmira Nushi
Soheil Feizi
Daniela Massiceti
ArXiv (abs)PDFHTML

Papers citing "Understanding Information Storage and Transfer in Multi-modal Large Language Models"

19 / 19 papers shown
Title
Head Pursuit: Probing Attention Specialization in Multimodal Transformers
Head Pursuit: Probing Attention Specialization in Multimodal Transformers
Lorenzo Basile
Valentino Maiorca
Diego Doimo
Francesco Locatello
Alberto Cazzaniga
85
0
0
24 Oct 2025
Map the Flow: Revealing Hidden Pathways of Information in VideoLLMs
Map the Flow: Revealing Hidden Pathways of Information in VideoLLMs
Minji Kim
Taekyung Kim
Bohyung Han
75
0
0
15 Oct 2025
Interpret, prune and distill Donut : towards lightweight VLMs for VQA on document
Interpret, prune and distill Donut : towards lightweight VLMs for VQA on document
Adnan Ben Mansour
Ayoub Karine
D. Naccache
80
0
0
30 Sep 2025
Skip-It? Theoretical Conditions for Layer Skipping in Vision-Language Models
Skip-It? Theoretical Conditions for Layer Skipping in Vision-Language Models
Max Hartman
Vidhata Arjun Jayaraman
Moulik Choraria
Akhil Bhimaraju
Lav Varshney
VLM
282
0
0
29 Sep 2025
Uncovering Grounding IDs: How External Cues Shape Multimodal Binding
Uncovering Grounding IDs: How External Cues Shape Multimodal Binding
Hosein Hasani
Amirmohammad Izadi
Fatemeh Askari
Mobin Bagherian
Sadegh Mohammadian
Mohammad Izadi
M. Baghshah
228
0
0
28 Sep 2025
REMA: A Unified Reasoning Manifold Framework for Interpreting Large Language Model
REMA: A Unified Reasoning Manifold Framework for Interpreting Large Language Model
Bo Li
Guanzhi Deng
Ronghao Chen
Junrong Yue
Shuo Zhang
Qinghua Zhao
Linqi Song
Lijie Wen
LRM
85
0
0
26 Sep 2025
Reading Images Like Texts: Sequential Image Understanding in Vision-Language Models
Reading Images Like Texts: Sequential Image Understanding in Vision-Language Models
Yueyan Li
Chenggong Zhao
Zeyuan Zang
Caixia Yuan
Xiaojie Wang
VLM
80
0
0
23 Sep 2025
Never Compromise to Vulnerabilities: A Comprehensive Survey on AI Governance
Never Compromise to Vulnerabilities: A Comprehensive Survey on AI Governance
Yuchu Jiang
Jian Zhao
Yuchen Yuan
Tianle Zhang
Yao Huang
...
Ya Zhang
Shuicheng Yan
Chi Zhang
Z. He
Xuelong Li
SILM
382
2
0
12 Aug 2025
FPEdit: Robust LLM Fingerprinting through Localized Parameter Editing
FPEdit: Robust LLM Fingerprinting through Localized Parameter Editing
Shida Wang
Chaohu Liu
Yubo Wang
Linli Xu
KELM
188
3
0
04 Aug 2025
Visual symbolic mechanisms: Emergent symbol processing in vision language models
Visual symbolic mechanisms: Emergent symbol processing in vision language models
Rim Assouel
Declan Campbell
Taylor Webb
118
2
0
18 Jun 2025
Same Task, Different Circuits: Disentangling Modality-Specific Mechanisms in VLMs
Yaniv Nikankin
Dana Arad
Yossi Gandelsman
Yonatan Belinkov
264
5
0
10 Jun 2025
Skip-Vision: Efficient and Scalable Acceleration of Vision-Language Models via Adaptive Token Skipping
Skip-Vision: Efficient and Scalable Acceleration of Vision-Language Models via Adaptive Token Skipping
Weili Zeng
Ziyuan Huang
Kaixiang Ji
Manwen Liao
VLM
530
3
0
26 Mar 2025
See What You Are Told: Visual Attention Sink in Large Multimodal ModelsInternational Conference on Learning Representations (ICLR), 2025
Seil Kang
Jinyeong Kim
Junhyeok Kim
Seong Jae Hwang
VLM
278
42
0
05 Mar 2025
Visual Attention Never Fades: Selective Progressive Attention ReCalibration for Detailed Image Captioning in Multimodal Large Language Models
Visual Attention Never Fades: Selective Progressive Attention ReCalibration for Detailed Image Captioning in Multimodal Large Language Models
Mingi Jung
Saehuyng Lee
Eunji Kim
Sungroh Yoon
904
6
0
03 Feb 2025
Performance Gap in Entity Knowledge Extraction Across Modalities in Vision Language Models
Performance Gap in Entity Knowledge Extraction Across Modalities in Vision Language ModelsAnnual Meeting of the Association for Computational Linguistics (ACL), 2024
Ido Cohen
Daniela Gottesman
Mor Geva
Raja Giryes
VLM
409
4
1
18 Dec 2024
Lifelong Knowledge Editing for Vision Language Models with Low-Rank Mixture-of-Experts
Lifelong Knowledge Editing for Vision Language Models with Low-Rank Mixture-of-ExpertsComputer Vision and Pattern Recognition (CVPR), 2024
Qizhou Chen
Chengyu Wang
Dakan Wang
Taolin Zhang
Wangyue Li
Xiaofeng He
KELM
277
5
0
23 Nov 2024
Understanding Multimodal LLMs: the Mechanistic Interpretability of Llava in Visual Question Answering
Zeping Yu
Sophia Ananiadou
961
8
0
17 Nov 2024
From Pixels to Prose: Advancing Multi-Modal Language Models for Remote Sensing
From Pixels to Prose: Advancing Multi-Modal Language Models for Remote Sensing
Xingwu Sun
Cheng Fei
Charles Zhang
Fei Jin
Qian Niu
...
Pohsun Feng
Ziqian Bi
Ming Liu
Yujiao Shi
Yichao Zhang
230
2
0
05 Nov 2024
Towards Interpreting Visual Information Processing in Vision-Language Models
Towards Interpreting Visual Information Processing in Vision-Language ModelsInternational Conference on Learning Representations (ICLR), 2024
Philip Quirke
Luke Ong
Juil Sock
Mor Geva
David M. Krueger
Fazl Barez
450
45
0
09 Oct 2024
1