88
0

Diffusing DeBias: Synthetic Bias Amplification for Model Debiasing

Abstract

Deep learning model effectiveness in classification tasks is often challenged by the quality and quantity of training data whenever they are affected by strong spurious correlations between specific attributes and target labels. This results in a form of bias affecting training data, which typically leads to unrecoverable weak generalization in prediction. This paper aims at facing this problem by leveraging bias amplification with generated synthetic data: we introduce Diffusing DeBias (DDB), a novel approach acting as a plug-in for common methods of unsupervised model debiasing exploiting the inherent bias-learning tendency of diffusion models in data generation. Specifically, our approach adopts conditional diffusion models to generate synthetic bias-aligned images, which replace the original training set for learning an effective bias amplifier model that we subsequently incorporate into an end-to-end and a two-step unsupervised debiasing approach. By tackling the fundamental issue of bias-conflicting training samples memorization in learning auxiliary models, typical of this type of techniques, our proposed method beats current state-of-the-art in multiple benchmark datasets, demonstrating its potential as a versatile and effective tool for tackling bias in deep learning models.

View on arXiv
@article{ciranni2025_2502.09564,
  title={ Diffusing DeBias: Synthetic Bias Amplification for Model Debiasing },
  author={ Massimiliano Ciranni and Vito Paolo Pastore and Roberto Di Via and Enzo Tartaglione and Francesca Odone and Vittorio Murino },
  journal={arXiv preprint arXiv:2502.09564},
  year={ 2025 }
}
Comments on this paper