ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2503.12801
39
0

BLIA: Detect model memorization in binary classification model through passive Label Inference attack

17 March 2025
Mohammad Wahiduzzaman Khan
Sheng Chen
Ilya Mironov
Leizhen Zhang
Rabib Noor
ArXivPDFHTML
Abstract

Model memorization has implications for both the generalization capacity of machine learning models and the privacy of their training data. This paper investigates label memorization in binary classification models through two novel passive label inference attacks (BLIA). These attacks operate passively, relying solely on the outputs of pre-trained models, such as confidence scores and log-loss values, without interacting with or modifying the training process. By intentionally flipping 50% of the labels in controlled subsets, termed "canaries," we evaluate the extent of label memorization under two conditions: models trained without label differential privacy (Label-DP) and those trained with randomized response-based Label-DP. Despite the application of varying degrees of Label-DP, the proposed attacks consistently achieve success rates exceeding 50%, surpassing the baseline of random guessing and conclusively demonstrating that models memorize training labels, even when these labels are deliberately uncorrelated with the features.

View on arXiv
@article{khan2025_2503.12801,
  title={ BLIA: Detect model memorization in binary classification model through passive Label Inference attack },
  author={ Mohammad Wahiduzzaman Khan and Sheng Chen and Ilya Mironov and Leizhen Zhang and Rabib Noor },
  journal={arXiv preprint arXiv:2503.12801},
  year={ 2025 }
}
Comments on this paper