Communities
Connect sessions
AI calendar
Organizations
Contact Sales
Search
Open menu
Home
Papers
2203.03929
Cited By
v1
v2 (latest)
Quantifying Privacy Risks of Masked Language Models Using Membership Inference Attacks
8 March 2022
Fatemehsadat Mireshghallah
Kartik Goyal
Archit Uniyal
Taylor Berg-Kirkpatrick
Reza Shokri
MIALM
Re-assign community
ArXiv (abs)
PDF
HTML
Papers citing
"Quantifying Privacy Risks of Masked Language Models Using Membership Inference Attacks"
50 / 122 papers shown
Title
PaCoST: Paired Confidence Significance Testing for Benchmark Contamination Detection in Large Language Models
Huixuan Zhang
Yun Lin
Xiaojun Wan
176
1
0
26 Jun 2024
ReCaLL: Membership Inference via Relative Conditional Log-Likelihoods
Roy Xie
Junlin Wang
Ruomin Huang
Minxing Zhang
Rong Ge
Jian Pei
Neil Zhenqiang Gong
Bhuwan Dhingra
MIALM
214
23
0
23 Jun 2024
Uncovering Latent Memories: Assessing Data Leakage and Memorization Patterns in Frontier AI Models
Sunny Duan
Mikail Khona
Abhiram Iyer
Rylan Schaeffer
Ila R Fiete
180
4
0
20 Jun 2024
Semantic Membership Inference Attack against Large Language Models
Hamid Mozaffari
Virendra J. Marathe
MIALM
128
6
0
14 Jun 2024
Unique Security and Privacy Threats of Large Language Models: A Comprehensive Survey
Shang Wang
Tianqing Zhu
B. Liu
Ming Ding
Dayong Ye
Dayong Ye
Wanlei Zhou
PILM
201
2
0
12 Jun 2024
Deconstructing The Ethics of Large Language Models from Long-standing Issues to New-emerging Dilemmas
Chengyuan Deng
Yiqun Duan
Xin Jin
Heng Chang
Yijun Tian
...
Kuofeng Gao
Sihong He
Jun Zhuang
Lu Cheng
Haohan Wang
AILaw
125
28
0
08 Jun 2024
Reconstructing training data from document understanding models
Jérémie Dentan
Arnaud Paran
A. Shabou
AAML
SyDa
120
2
0
05 Jun 2024
Probing Language Models for Pre-training Data Detection
Zhenhua Liu
Tong Zhu
Chuanyuan Tan
Haonan Lu
Bing Liu
Wenliang Chen
114
18
0
03 Jun 2024
ConStat: Performance-Based Contamination Detection in Large Language Models
Jasper Dekoninck
Mark Niklas Muller
Martin Vechev
75
9
0
25 May 2024
Better Membership Inference Privacy Measurement through Discrepancy
Ruihan Wu
Pengrun Huang
Kamalika Chaudhuri
MIACV
93
1
0
24 May 2024
Data Contamination Calibration for Black-box LLMs
Wen-song Ye
Jiaqi Hu
Liyao Li
Haobo Wang
Gang Chen
Junbo Zhao
84
13
0
20 May 2024
Many-Shot Regurgitation (MSR) Prompting
Shashank Sonkar
Richard G. Baraniuk
AAML
75
1
0
13 May 2024
Elephants Never Forget: Memorization and Learning of Tabular Data in Large Language Models
Sebastian Bordt
Harsha Nori
Vanessa Rodrigues
Besmira Nushi
Rich Caruana
128
22
0
09 Apr 2024
Min-K%++: Improved Baseline for Detecting Pre-Training Data from Large Language Models
Jingyang Zhang
Jingwei Sun
Eric C. Yeats
Ouyang Yang
Martin Kuo
Jianyi Zhang
Hao Frank Yang
Hai "Helen" Li
253
67
0
03 Apr 2024
Don't Listen To Me: Understanding and Exploring Jailbreak Prompts of Large Language Models
Zhiyuan Yu
Xiaogeng Liu
Shunning Liang
Zach Cameron
Chaowei Xiao
Ning Zhang
126
67
0
26 Mar 2024
PreCurious: How Innocent Pre-Trained Language Models Turn into Privacy Traps
Ruixuan Liu
Tianhao Wang
Yang Cao
Li Xiong
AAML
SILM
246
24
0
14 Mar 2024
Second-Order Information Matters: Revisiting Machine Unlearning for Large Language Models
Kang Gu
Md Rafi Ur Rashid
Najrin Sultana
Shagufta Mehnaz
MU
120
12
0
13 Mar 2024
On Protecting the Data Privacy of Large Language Models (LLMs): A Survey
Biwei Yan
Kun Li
Minghui Xu
Yueyan Dong
Yue Zhang
Zhaochun Ren
Xiuzhen Cheng
AILaw
PILM
211
115
0
08 Mar 2024
Amplifying Training Data Exposure through Fine-Tuning with Pseudo-Labeled Memberships
Myung Gyo Oh
Hong Eun Ahn
L. Park
T.-H. Kwon
MIALM
AAML
132
0
0
19 Feb 2024
DE-COP: Detecting Copyrighted Content in Language Models Training Data
André V. Duarte
Xuandong Zhao
Arlindo L. Oliveira
Lei Li
160
54
0
15 Feb 2024
Copyright Traps for Large Language Models
Matthieu Meeus
Igor Shilov
Manuel Faysse
Yves-Alexandre de Montjoye
159
27
0
14 Feb 2024
Do Membership Inference Attacks Work on Large Language Models?
Michael Duan
Anshuman Suri
Niloofar Mireshghallah
Sewon Min
Weijia Shi
Luke Zettlemoyer
Yulia Tsvetkov
Yejin Choi
David Evans
Hanna Hajishirzi
MIALM
174
130
0
12 Feb 2024
Mitigating Privacy Risk in Membership Inference by Convex-Concave Loss
Zhenlong Liu
Lei Feng
Huiping Zhuang
Xiaofeng Cao
Jianguo Huang
95
3
0
08 Feb 2024
Evading Data Contamination Detection for Language Models is (too) Easy
Jasper Dekoninck
Mark Niklas Muller
Maximilian Baader
Marc Fischer
Martin Vechev
229
23
0
05 Feb 2024
Machine Unlearning in Large Language Models
Kongyang Chen
Zixin Wang
Bing Mi
Waixi Liu
Shaowei Wang
Xiaojun Ren
Jiaxing Shen
MU
100
18
0
03 Feb 2024
Security and Privacy Challenges of Large Language Models: A Survey
B. Das
M. H. Amini
Yanzhao Wu
PILM
ELM
203
214
0
30 Jan 2024
Fortifying Ethical Boundaries in AI: Advanced Strategies for Enhancing Security in Large Language Models
Yunhong He
Jianling Qiu
Wei Zhang
Zhe Yuan
82
3
0
27 Jan 2024
Investigating Data Contamination for Pre-training Language Models
Minhao Jiang
Ken Ziyu Liu
Ming Zhong
Rylan Schaeffer
Siru Ouyang
Jiawei Han
Sanmi Koyejo
140
80
0
11 Jan 2024
Private Fine-tuning of Large Language Models with Zeroth-order Optimization
Xinyu Tang
Ashwinee Panda
Milad Nasr
Saeed Mahloujifar
Prateek Mittal
304
31
0
09 Jan 2024
Digger: Detecting Copyright Content Mis-usage in Large Language Model Training
Haodong Li
Gelei Deng
Yi Liu
Kailong Wang
Yuekang Li
Tianwei Zhang
Yang Liu
Guoai Xu
Guosheng Xu
Haoyu Wang
128
28
0
01 Jan 2024
Traces of Memorisation in Large Language Models for Code
Ali Al-Kaswan
Maliheh Izadi
Arie van Deursen
ELM
97
26
0
18 Dec 2023
A Comprehensive Survey of Attack Techniques, Implementation, and Mitigation Strategies in Large Language Models
Aysan Esmradi
Daniel Wankit Yip
C. Chan
AAML
115
16
0
18 Dec 2023
A Survey on Large Language Model (LLM) Security and Privacy: The Good, the Bad, and the Ugly
Yifan Yao
Jinhao Duan
Kaidi Xu
Yuanfang Cai
Eric Sun
Yue Zhang
PILM
ELM
252
743
0
04 Dec 2023
DP-OPT: Make Large Language Model Your Privacy-Preserving Prompt Engineer
Junyuan Hong
Jiachen T. Wang
Chenhui Zhang
Zhangheng Li
Yue Liu
Zhangyang Wang
195
48
0
27 Nov 2023
Fuse to Forget: Bias Reduction and Selective Memorization through Model Fusion
Kerem Zaman
Leshem Choshen
Shashank Srivastava
KELM
MoMe
132
13
0
13 Nov 2023
Practical Membership Inference Attacks against Fine-tuned Large Language Models via Self-prompt Calibration
Wenjie Fu
Huandong Wang
Chen Gao
Guanghua Liu
Yong Li
Tao Jiang
MIALM
135
62
0
10 Nov 2023
Detecting Pretraining Data from Large Language Models
Weijia Shi
Anirudh Ajith
Mengzhou Xia
Yangsibo Huang
Daogao Liu
Terra Blevins
Danqi Chen
Luke Zettlemoyer
MIALM
180
252
0
25 Oct 2023
FLTrojan: Privacy Leakage Attacks against Federated Language Models Through Selective Weight Tampering
Md Rafi Ur Rashid
Vishnu Asutosh Dasu
Kang Gu
Najrin Sultana
Shagufta Mehnaz
AAML
FedML
262
12
0
24 Oct 2023
Assessing Privacy Risks in Language Models: A Case Study on Summarization Tasks
Ruixiang Tang
Gord Lueck
Rodolfo Quispe
Huseyin A. Inan
Janardhan Kulkarni
Helen Zhou
119
8
0
20 Oct 2023
Last One Standing: A Comparative Analysis of Security and Privacy of Soft Prompt Tuning, LoRA, and In-Context Learning
Rui Wen
Tianhao Wang
Michael Backes
Yang Zhang
Ahmed Salem
AAML
115
14
0
17 Oct 2023
Privacy in Large Language Models: Attacks, Defenses and Future Directions
Haoran Li
Yulin Chen
Jinglong Luo
Yan Kang
Xiaojin Zhang
Qi Hu
Chunkit Chan
Yangqiu Song
PILM
177
54
0
16 Oct 2023
User Inference Attacks on Large Language Models
Nikhil Kandpal
Krishna Pillutla
Alina Oprea
Peter Kairouz
Christopher A. Choquette-Choo
Zheng Xu
SILM
AAML
157
28
0
13 Oct 2023
Improved Membership Inference Attacks Against Language Classification Models
Shlomit Shachor
N. Razinkov
Abigail Goldsteen
131
7
0
11 Oct 2023
Regulation and NLP (RegNLP): Taming Large Language Models
Catalina Goanta
Nikolaos Aletras
Ilias Chalkidis
S. Ranchordas
Gerasimos Spanakis
AILaw
72
4
0
09 Oct 2023
Gotcha! This Model Uses My Code! Evaluating Membership Leakage Risks in Code Models
Zhou Yang
Zhipeng Zhao
Chenyu Wang
Jieke Shi
Dongsum Kim
Donggyun Han
David Lo
SILM
AAML
MIACV
157
13
0
02 Oct 2023
Recent Advances of Differential Privacy in Centralized Deep Learning: A Systematic Survey
Lea Demelius
Roman Kern
Andreas Trügler
SyDa
FedML
125
15
0
28 Sep 2023
Identifying and Mitigating Privacy Risks Stemming from Language Models: A Survey
Victoria Smith
Ali Shahin Shamsabadi
Carolyn Ashurst
Adrian Weller
PILM
183
35
0
27 Sep 2023
Large Language Model Alignment: A Survey
Shangda Wu
Renren Jin
Yufei Huang
Chuang Liu
Weilong Dong
Zishan Guo
Xinwei Wu
Yan Liu
Deyi Xiong
LM&MA
152
241
0
26 Sep 2023
SLMIA-SR: Speaker-Level Membership Inference Attacks against Speaker Recognition Systems
Guangke Chen
Yedi Zhang
Fu Song
121
11
0
14 Sep 2023
Jais and Jais-chat: Arabic-Centric Foundation and Instruction-Tuned Open Generative Large Language Models
Neha Sengupta
Sunil Kumar Sahu
Bokang Jia
Satheesh Katipomu
Haonan Li
...
A. Jackson
Hector Xuguang Ren
Preslav Nakov
Timothy Baldwin
Eric P. Xing
LRM
148
49
0
30 Aug 2023
Previous
1
2
3
Next