Communities
Connect sessions
AI calendar
Organizations
Contact Sales
Search
Open menu
Home
Papers
2203.03929
Cited By
v1
v2 (latest)
Quantifying Privacy Risks of Masked Language Models Using Membership Inference Attacks
8 March 2022
Fatemehsadat Mireshghallah
Kartik Goyal
Archit Uniyal
Taylor Berg-Kirkpatrick
Reza Shokri
MIALM
Re-assign community
ArXiv (abs)
PDF
HTML
Papers citing
"Quantifying Privacy Risks of Masked Language Models Using Membership Inference Attacks"
50 / 122 papers shown
Title
Enterprise AI Must Enforce Participant-Aware Access Control
Shashank Shreedhar Bhatt
Tanmay Rajore
Khushboo Aggarwal
Ganesh Ananthanarayanan
Ranveer Chandra
...
Emre Kiciman
Sumit Kumar Pandey
Srinath T. V. Setty
Rahul Sharma
Teijia Zhao
AAML
SILM
8
0
0
18 Sep 2025
The Sum Leaks More Than Its Parts: Compositional Privacy Risks and Mitigations in Multi-Agent Collaboration
Vaidehi Patil
Elias Stengel-Eskin
Mohit Bansal
20
0
0
16 Sep 2025
Beyond PII: How Users Attempt to Estimate and Mitigate Implicit LLM Inference
Synthia Wang
Sai Teja Peddinti
Nina Taft
Nick Feamster
SILM
PILM
4
1
0
15 Sep 2025
LLM in the Middle: A Systematic Review of Threats and Mitigations to Real-World LLM-based Systems
Vitor Hugo Galhardo Moia
Igor Jochem Sanz
Gabriel Antonio Fontes Rebello
Rodrigo Duarte de Meneses
Briland Hitaj
Ulf Lindqvist
52
0
0
12 Sep 2025
Active Membership Inference Test (aMINT): Enhancing Model Auditability with Multi-Task Learning
Daniel DeAlcala
Aythami Morales
Julian Fierrez
Gonzalo Mancera
Ruben Tolosana
Javier-Ortega Garcia
MLAU
56
0
0
09 Sep 2025
Privacy Risks in Time Series Forecasting: User- and Record-Level Membership Inference
Nicolas Johansson
Tobias Olsson
Daniel Nilsson
Johan Östman
Fazeleh Hoseini
AI4TS
16
0
0
04 Sep 2025
Ensembling Membership Inference Attacks Against Tabular Generative Models
Joshua Ward
Yuxuan Yang
Chi-Hua Wang
Guang Cheng
0
0
0
02 Sep 2025
Never Compromise to Vulnerabilities: A Comprehensive Survey on AI Governance
Yuchu Jiang
Jian Zhao
Yuchen Yuan
Tianle Zhang
Yao Huang
...
Ya Zhang
Shuicheng Yan
Chi Zhang
Z. He
Xuelong Li
SILM
106
1
0
12 Aug 2025
Win-k: Improved Membership Inference Attacks on Small Language Models
Roya Arkhmammadova
Hosein Madadi Tamar
M. Emre Gursoy
37
0
0
02 Aug 2025
Uncovering Gradient Inversion Risks in Practical Language Model Training
Xinguo Feng
Zhongkui Ma
Zihan Wang
Eu Joe Chegne
Mengyao Ma
Alsharif Abuadbba
Guangdong Bai
86
5
0
28 Jul 2025
LoRA-Leak: Membership Inference Attacks Against LoRA Fine-tuned Language Models
Delong Ran
Xinlei He
Tianshuo Cong
Anyu Wang
Cunliang Kong
Xiaoyun Wang
MIALM
PILM
89
0
1
24 Jul 2025
Winter Soldier: Backdooring Language Models at Pre-Training with Indirect Data Poisoning
Wassim Bouaziz
Mathurin Videau
Nicolas Usunier
El-Mahdi El-Mhamdi
87
0
0
17 Jun 2025
Private Memorization Editing: Turning Memorization into a Defense to Strengthen Data Privacy in Large Language Models
Elena Sofia Ruzzetti
Giancarlo A. Xompero
Davide Venditti
Fabio Massimo Zanzotto
KELM
PILM
101
0
0
09 Jun 2025
HACo-Det: A Study Towards Fine-Grained Machine-Generated Text Detection under Human-AI Coauthoring
Zhixiong Su
Yichen Wang
Herun Wan
Zhaohan Zhang
Minnan Luo
DeLMO
133
2
0
03 Jun 2025
Permissioned LLMs: Enforcing Access Control in Large Language Models
Bargav Jayaraman
Virendra J. Marathe
Hamid Mozaffari
William F. Shen
Krishnaram Kenthapadi
138
0
0
28 May 2025
Strong Membership Inference Attacks on Massive Datasets and (Moderately) Large Language Models
Jamie Hayes
Ilia Shumailov
Christopher A. Choquette-Choo
Matthew Jagielski
G. Kaissis
...
Matthieu Meeus
Yves-Alexandre de Montjoye
Franziska Boenisch
Adam Dziedzic
A. Feder Cooper
116
7
0
24 May 2025
Fragments to Facts: Partial-Information Fragment Inference from LLMs
Lucas Rosenblatt
Bin Han
Robert Wolfe
Bill Howe
AAML
108
0
0
20 May 2025
Automatic Calibration for Membership Inference Attack on Large Language Models
Saleh Zare Zade
Yao Qiang
Xiangyu Zhou
Hui Zhu
Mohammad Amin Roshani
Prashant Khanduri
Dongxiao Zhu
144
2
0
06 May 2025
Membership Inference Attacks on Large-Scale Models: A Survey
Hengyu Wu
Yang Cao
MIALM
297
4
0
25 Mar 2025
Empirical Calibration and Metric Differential Privacy in Language Models
Pedro Faustini
Natasha Fernandes
Annabelle McIver
Mark Dras
144
0
0
18 Mar 2025
Efficient Membership Inference Attacks by Bayesian Neural Network
Zhenlong Liu
Wenyu Jiang
Feng Zhou
Jianguo Huang
MIALM
166
1
0
10 Mar 2025
Is My Text in Your AI Model? Gradient-based Membership Inference Test applied to LLMs
Gonzalo Mancera
Daniel DeAlcala
Julian Fierrez
Ruben Tolosana
Aythami Morales
192
5
0
10 Mar 2025
Mitigating Memorization in LLMs using Activation Steering
Manan Suri
Nishit Anand
Amisha Bhaskar
LLMSV
154
4
0
08 Mar 2025
Watch Out Your Album! On the Inadvertent Privacy Memorization in Multi-Modal Large Language Models
Tianjie Ju
Yi Hua
Hao Fei
Zhenyu Shao
Yubin Zheng
Haodong Zhao
Yang Deng
Wynne Hsu
Zhuosheng Zhang
Gongshen Liu
189
1
0
03 Mar 2025
Towards Label-Only Membership Inference Attack against Pre-trained Large Language Models
Yu He
Boheng Li
Lu Liu
Zhongjie Ba
Wei Dong
Yiming Li
Zhan Qin
Kui Ren
Chong Chen
MIALM
212
9
0
26 Feb 2025
R.R.: Unveiling LLM Training Privacy through Recollection and Ranking
Wenlong Meng
Zhenyuan Guo
Lenan Wu
Chen Gong
Wenyan Liu
Weixian Li
Chengkun Wei
Wenzhi Chen
PILM
129
3
0
18 Feb 2025
Synthetic Data Privacy Metrics
Amy Steier
Lipika Ramaswamy
Andre Manoel
Alexa Haushalter
139
3
0
08 Jan 2025
Efficient and Private: Memorisation under differentially private parameter-efficient fine-tuning in language models
Olivia Ma
Jonathan Passerat-Palmbach
Dmitrii Usynin
176
1
0
24 Nov 2024
Membership Inference Attack against Long-Context Large Language Models
Zixiong Wang
Gaoyang Liu
Yang Yang
Chen Wang
218
1
0
18 Nov 2024
IntentGPT: Few-shot Intent Discovery with Large Language Models
Juan A. Rodriguez
Nicholas Botzer
David Vazquez
Christopher Pal
M. Pedersoli
I. Laradji
VLM
205
6
0
16 Nov 2024
Unlearning in- vs. out-of-distribution data in LLMs under gradient-based method
Teodora Baluta
Pascal Lamblin
Daniel Tarlow
Fabian Pedregosa
Gintare Karolina Dziugaite
MU
94
4
0
07 Nov 2024
Evaluation data contamination in LLMs: how do we measure it and (when) does it matter?
Aaditya K. Singh
Muhammed Yusuf Kocyigit
Andrew Poulton
David Esiobu
Maria Lomeli
Gergely Szilvasy
Dieuwke Hupkes
114
16
0
06 Nov 2024
Does Data Contamination Detection Work (Well) for LLMs? A Survey and Evaluation on Detection Assumptions
Yujuan Fu
Özlem Uzuner
Meliha Yetisgen
Fei Xia
189
12
0
24 Oct 2024
Reconstruction of Differentially Private Text Sanitization via Large Language Models
Shuchao Pang
Zhigang Lu
Jian Shu
Peng Fu
Yongbin Zhou
Minhui Xue
AAML
206
5
0
16 Oct 2024
Detecting Training Data of Large Language Models via Expectation Maximization
Gyuwan Kim
Yang Li
Evangelia Spiliopoulou
Jie Ma
Miguel Ballesteros
William Yang Wang
MIALM
360
7
2
10 Oct 2024
Defending Membership Inference Attacks via Privacy-aware Sparsity Tuning
Qiang Hu
Hengxiang Zhang
Jianguo Huang
158
2
0
09 Oct 2024
Undesirable Memorization in Large Language Models: A Survey
Ali Satvaty
Suzan Verberne
Fatih Turkmen
ELM
PILM
272
13
0
03 Oct 2024
Extracting Memorized Training Data via Decomposition
Ellen Su
Anu Vellore
Amy Chang
Raffaele Mura
Blaine Nelson
Paul Kassianik
Amin Karbasi
96
3
0
18 Sep 2024
Generated Data with Fake Privacy: Hidden Dangers of Fine-tuning Large Language Models on Generated Data
Atilla Akkus
Mingjie Li
Junjie Chu
Junjie Chu
Michael Backes
Sinem Sav
Sinem Sav
SILM
SyDa
188
7
0
12 Sep 2024
Context-Aware Membership Inference Attacks against Pre-trained Large Language Models
Hongyan Chang
Ali Shahin Shamsabadi
Kleomenis Katevas
Hamed Haddadi
Reza Shokri
MIALM
209
9
0
11 Sep 2024
Con-ReCall: Detecting Pre-training Data in LLMs via Contrastive Decoding
Cheng Wang
Yiwei Wang
Bryan Hooi
Yujun Cai
Nanyun Peng
Kai-Wei Chang
231
7
0
05 Sep 2024
Recent Advances in Attack and Defense Approaches of Large Language Models
Jing Cui
Yishi Xu
Zhewei Huang
Shuchang Zhou
Jianbin Jiao
Junge Zhang
PILM
AAML
217
7
0
05 Sep 2024
LLM-PBE: Assessing Data Privacy in Large Language Models
Qinbin Li
Junyuan Hong
Chulin Xie
Jeffrey Tan
Rachel Xin
...
Dan Hendrycks
Zhangyang Wang
Bo Li
Bingsheng He
Dawn Song
ELM
PILM
172
34
0
23 Aug 2024
Inside the Black Box: Detecting Data Leakage in Pre-trained Language Encoders
Yuan Xin
Hui Yuan
Ning Yu
Dingfan Chen
Mario Fritz
Michael Backes
Yang Zhang
PILM
MIACV
131
2
0
20 Aug 2024
Adaptive Pre-training Data Detection for Large Language Models via Surprising Tokens
Anqi Zhang
Chaofeng Wu
149
6
0
30 Jul 2024
Data Mixture Inference: What do BPE Tokenizers Reveal about their Training Data?
J. Hayase
Alisa Liu
Yejin Choi
Sewoong Oh
Noah A. Smith
200
17
0
23 Jul 2024
Operationalizing a Threat Model for Red-Teaming Large Language Models (LLMs)
Apurv Verma
Satyapriya Krishna
Sebastian Gehrmann
Madhavan Seshadri
Anu Pradhan
Tom Ault
Leslie Barrett
David Rabinowitz
John Doucette
Nhathai Phan
191
32
0
20 Jul 2024
Was it Slander? Towards Exact Inversion of Generative Language Models
Adrians Skapars
Edoardo Manino
Youcheng Sun
Lucas C. Cordeiro
111
4
0
10 Jul 2024
ObfuscaTune: Obfuscated Offsite Fine-tuning and Inference of Proprietary LLMs on Private Datasets
Ahmed Frikha
Nassim Walha
Ricardo Mendes
Krishna Kanth Nakka
Xue Jiang
Xuebing Zhou
222
4
0
03 Jul 2024
A Method to Facilitate Membership Inference Attacks in Deep Learning Models
Zitao Chen
Karthik Pattabiraman
MIACV
MLAU
AAML
MIALM
169
2
0
02 Jul 2024
1
2
3
Next