ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2411.03823
41
0

Both Text and Images Leaked! A Systematic Analysis of Multimodal LLM Data Contamination

6 November 2024
D. Song
Sicheng Lai
Shunian Chen
Lichao Sun
Benyou Wang
ArXivPDFHTML
Abstract

The rapid progression of multimodal large language models (MLLMs) has demonstrated superior performance on various multimodal benchmarks. However, the issue of data contamination during training creates challenges in performance evaluation and comparison. While numerous methods exist for detecting models' contamination in large language models (LLMs), they are less effective for MLLMs due to their various modalities and multiple training phases. In this study, we introduce a multimodal data contamination detection framework, MM-Detect, designed for MLLMs. Our experimental results indicate that MM-Detect is quite effective and sensitive in identifying varying degrees of contamination, and can highlight significant performance improvements due to the leakage of multimodal benchmark training sets. Furthermore, we explore whether the contamination originates from the base LLMs used by MLLMs or the multimodal training phase, providing new insights into the stages at which contamination may be introduced.

View on arXiv
@article{song2025_2411.03823,
  title={ Both Text and Images Leaked! A Systematic Analysis of Multimodal LLM Data Contamination },
  author={ Dingjie Song and Sicheng Lai and Shunian Chen and Lichao Sun and Benyou Wang },
  journal={arXiv preprint arXiv:2411.03823},
  year={ 2025 }
}
Comments on this paper