Conceptual Contrastive Edits in Textual and Vision-Language Retrieval
Abstract
As deep learning models grow in complexity, achieving model-agnostic interpretability becomes increasingly vital. In this work, we employ post-hoc conceptual contrastive edits to expose noteworthy patterns and biases imprinted in representations of retrieval models. We systematically design optimal and controllable contrastive interventions targeting various parts of speech, and effectively apply them to explain both linguistic and visiolinguistic pre-trained models in a black-box manner. Additionally, we introduce a novel metric to assess the per-word impact of contrastive interventions on model outcomes, providing a comprehensive evaluation of each intervention's effectiveness.
View on arXiv@article{lymperaiou2025_2503.01914, title={ Conceptual Contrastive Edits in Textual and Vision-Language Retrieval }, author={ Maria Lymperaiou and Giorgos Stamou }, journal={arXiv preprint arXiv:2503.01914}, year={ 2025 } }
Comments on this paper