Is My Text in Your AI Model? Gradient-based Membership Inference Test applied to LLMs
This work adapts and studies the gradient-based Membership Inference Test (gMINT) to the classification of text based on LLMs. MINT is a general approach intended to determine if given data was used for training machine learning models, and this work focuses on its application to the domain of Natural Language Processing. Using gradient-based analysis, the MINT model identifies whether particular data samples were included during the language model training phase, addressing growing concerns about data privacy in machine learning. The method was evaluated in seven Transformer-based models and six datasets comprising over 2.5 million sentences, focusing on text classification tasks. Experimental results demonstrate MINTs robustness, achieving AUC scores between 85% and 99%, depending on data size and model architecture. These findings highlight MINTs potential as a scalable and reliable tool for auditing machine learning models, ensuring transparency, safeguarding sensitive data, and fostering ethical compliance in the deployment of AI/NLP technologies.
View on arXiv@article{mancera2025_2503.07384, title={ Is My Text in Your AI Model? Gradient-based Membership Inference Test applied to LLMs }, author={ Gonzalo Mancera and Daniel DeAlcala and Julian Fierrez and Ruben Tolosana and Aythami Morales }, journal={arXiv preprint arXiv:2503.07384}, year={ 2025 } }