Sustainable transparency in Recommender Systems: Bayesian Ranking of Images for Explainability

Recommender Systems have become crucial in the modern world, commonly guiding users towards relevant content or products, and having a large influence over the decisions of users and citizens. However, ensuring transparency and user trust in these systems remains a challenge; personalized explanations have emerged as a solution, offering justifications for recommendations. Among the existing approaches for generating personalized explanations, using existing visual content created by users is a promising option to maximize transparency and user trust. State-of-the-art models that follow this approach, despite leveraging highly optimized architectures, employ surrogate learning tasks that do not efficiently model the objective of ranking images as explanations for a given recommendation; this leads to a suboptimal training process with high computational costs that may not be reduced without affecting model performance. This work presents BRIE, a novel model where we leverage Bayesian Pairwise Ranking to enhance the training process, allowing us to consistently outperform state-of-the-art models in six real-world datasets while reducing its model size by up to 64 times and its CO2 emissions by up to 75% in training and inference.
View on arXiv@article{paz-ruza2025_2308.01196, title={ Sustainable transparency in Recommender Systems: Bayesian Ranking of Images for Explainability }, author={ Jorge Paz-Ruza and Amparo Alonso-Betanzos and Berta Guijarro-Berdiñas and Brais Cancela and Carlos Eiras-Franco }, journal={arXiv preprint arXiv:2308.01196}, year={ 2025 } }