ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2505.15425
78
0
v1v2 (latest)

On the Robustness of Medical Vision-Language Models: Are they Truly Generalizable?

21 May 2025
Raza Imam
Rufael Marew
Mohammad Yaqub
    AAMLVLM
ArXiv (abs)PDFHTML
Main:15 Pages
15 Figures
Bibliography:2 Pages
6 Tables
Appendix:6 Pages
Abstract

Medical Vision-Language Models (MVLMs) have achieved par excellence generalization in medical image analysis, yet their performance under noisy, corrupted conditions remains largely untested. Clinical imaging is inherently susceptible to acquisition artifacts and noise; however, existing evaluations predominantly assess generally clean datasets, overlooking robustness -- i.e., the model's ability to perform under real-world distortions. To address this gap, we first introduce MediMeta-C, a corruption benchmark that systematically applies several perturbations across multiple medical imaging datasets. Combined with MedMNIST-C, this establishes a comprehensive robustness evaluation framework for MVLMs. We further propose RobustMedCLIP, a visual encoder adaptation of a pretrained MVLM that incorporates few-shot tuning to enhance resilience against corruptions. Through extensive experiments, we benchmark 5 major MVLMs across 5 medical imaging modalities, revealing that existing models exhibit severe degradation under corruption and struggle with domain-modality tradeoffs. Our findings highlight the necessity of diverse training and robust adaptation strategies, demonstrating that efficient low-rank adaptation when paired with few-shot tuning, improves robustness while preserving generalization across modalities.

View on arXiv
@article{imam2025_2505.15425,
  title={ On the Robustness of Medical Vision-Language Models: Are they Truly Generalizable? },
  author={ Raza Imam and Rufael Marew and Mohammad Yaqub },
  journal={arXiv preprint arXiv:2505.15425},
  year={ 2025 }
}
Comments on this paper