Large vision-language models (LVLMs) generally contain significantly more visual tokens than their textual counterparts, resulting in a considerable computational burden. Recent efforts have been made to tackle this issue by pruning visual tokens early within the language model. Most existing works use attention scores between text and visual tokens to assess the importance of visual tokens. However, in this study, we first analyze the text-visual attention in the language model and find that this score is not an ideal indicator for token pruning. Based on the analysis, We propose VisPruner, a plug-and-play method that utilizes visual cues for more effective token pruning in LVLMs. Specifically, we first use visual attention to select a limited number of significant tokens. Then, we remove duplicate tokens from the remaining ones based on their similarity. By retaining diverse tokens alongside the initially selected important tokens, we maximally preserve the visual information of the input image. Experimental results demonstrate that our VisPruner sustains strong performance across various VLM architectures and reduction ratios, significantly outperforming existing methods based on text-visual attention. Notably, without any training, VisPruner can reduce the FLOPs of LLaVA-1.5-7B by 91% and inference latency by 75%, while maintaining comparable performance. Our code is available atthis https URL.
View on arXiv@article{zhang2025_2412.01818, title={ Beyond Text-Visual Attention: Exploiting Visual Cues for Effective Token Pruning in VLMs }, author={ Qizhe Zhang and Aosong Cheng and Ming Lu and Renrui Zhang and Zhiyong Zhuo and Jiajun Cao and Shaobo Guo and Qi She and Shanghang Zhang }, journal={arXiv preprint arXiv:2412.01818}, year={ 2025 } }