ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2505.03153
19
0

Robust Fairness Vision-Language Learning for Medical Image Analysis

6 May 2025
Sparsh Bansal
Mingyang Wu
Xin Wang
S. Hu
    VLM
ArXivPDFHTML
Abstract

The advent of Vision-Language Models (VLMs) in medical image analysis has the potential to help process multimodal inputs and increase performance over traditional inference methods. However, when considering the domain in which these models will be implemented, fairness and robustness are important to ensure the model stays true for any patient. In this paper, we introduce a framework for ensuring robustness and fairness of VLM models. This framework modifies the loss function at training by identifying and adjusting faulty image-text pairs through a Dynamic Bad Pair Mining algorithm and also utilizing Sinkhorn distance to ensure the loss distributions of protected groups do not deviate from the total loss. Experimental testing of our framework shows up to a 8.6\% improvement when looking at equity-scaled AUC.

View on arXiv
@article{bansal2025_2505.03153,
  title={ Robust Fairness Vision-Language Learning for Medical Image Analysis },
  author={ Sparsh Bansal and Mingyang Wu and Xin Wang and Shu Hu },
  journal={arXiv preprint arXiv:2505.03153},
  year={ 2025 }
}
Comments on this paper