Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2004.00053
Cited By
Information Leakage in Embedding Models
31 March 2020
Congzheng Song
A. Raghunathan
MIACV
Re-assign community
ArXiv
PDF
HTML
Papers citing
"Information Leakage in Embedding Models"
31 / 31 papers shown
Title
Prompt Inversion Attack against Collaborative Inference of Large Language Models
Wenjie Qu
Yuguang Zhou
Yongji Wu
Tingsong Xiao
Binhang Yuan
Y. Li
Jiaheng Zhang
66
0
0
12 Mar 2025
Membership Inference Risks in Quantized Models: A Theoretical and Empirical Study
Eric Aubinais
Philippe Formont
Pablo Piantanida
Elisabeth Gassiat
38
0
0
10 Feb 2025
Top Ten Challenges Towards Agentic Neural Graph Databases
Jiaxin Bai
Z. Wang
Yukun Zhou
Hang Yin
WeiZhi Fei
...
Binhang Yuan
Wei Wang
Lei Chen
Xiaofang Zhou
Y. Song
52
0
0
24 Jan 2025
GRID: Protecting Training Graph from Link Stealing Attacks on GNN Models
Jiadong Lou
Xu Yuan
Rui Zhang
Xingliang Yuan
Neil Gong
N. Tzeng
AAML
33
1
0
19 Jan 2025
Navigating the Designs of Privacy-Preserving Fine-tuning for Large Language Models
Haonan Shi
Tu Ouyang
An Wang
31
0
0
08 Jan 2025
On the Vulnerability of Text Sanitization
Meng Tong
Kejiang Chen
Xiaojian Yuang
J. Liu
W. Zhang
Nenghai Yu
Jie Zhang
47
0
0
22 Oct 2024
Large Language Models are Easily Confused: A Quantitative Metric, Security Implications and Typological Analysis
Yiyi Chen
Qiongxiu Li
Russa Biswas
Johannes Bjerva
34
1
0
17 Oct 2024
A Different Level Text Protection Mechanism With Differential Privacy
Qingwen Fu
20
0
0
05 Sep 2024
Privacy Checklist: Privacy Violation Detection Grounding on Contextual Integrity Theory
Haoran Li
Wei Fan
Yulin Chen
Jiayang Cheng
Tianshu Chu
Xuebing Zhou
Peizhao Hu
Yangqiu Song
AILaw
37
2
0
19 Aug 2024
Transferable Embedding Inversion Attack: Uncovering Privacy Risks in Text Embeddings without Model Queries
Yu-Hsiang Huang
Yuche Tsai
Hsiang Hsiao
Hong-Yi Lin
Shou-De Lin
SILM
35
6
0
12 Jun 2024
Reconstructing training data from document understanding models
Jérémie Dentan
Arnaud Paran
A. Shabou
AAML
SyDa
34
1
0
05 Jun 2024
Membership Inference Attacks and Privacy in Topic Modeling
Nico Manzonelli
Wanrong Zhang
Salil P. Vadhan
19
1
0
07 Mar 2024
Fundamental Limits of Membership Inference Attacks on Machine Learning Models
Eric Aubinais
Elisabeth Gassiat
Pablo Piantanida
MIACV
35
2
0
20 Oct 2023
Text Embeddings Reveal (Almost) As Much As Text
John X. Morris
Volodymyr Kuleshov
Vitaly Shmatikov
Alexander M. Rush
RALM
24
89
0
10 Oct 2023
Ethicist: Targeted Training Data Extraction Through Loss Smoothed Soft Prompting and Calibrated Confidence Estimation
Zhexin Zhang
Jiaxin Wen
Minlie Huang
17
29
0
10 Jul 2023
Protecting User Privacy in Remote Conversational Systems: A Privacy-Preserving framework based on text sanitization
Zhigang Kan
Linbo Qiao
Hao Yu
Liwen Peng
Yifu Gao
Dongsheng Li
11
20
0
14 Jun 2023
Privacy-Preserving Prompt Tuning for Large Language Model Services
Yansong Li
Zhixing Tan
Yang Liu
SILM
VLM
41
63
0
10 May 2023
On the Adversarial Inversion of Deep Biometric Representations
Gioacchino Tangari
Shreesh Keskar
H. Asghar
Dali Kaafar
AAML
12
2
0
12 Apr 2023
Multi-step Jailbreaking Privacy Attacks on ChatGPT
Haoran Li
Dadi Guo
Wei Fan
Mingshi Xu
Jie Huang
Fanpu Meng
Yangqiu Song
SILM
19
317
0
11 Apr 2023
"Real Attackers Don't Compute Gradients": Bridging the Gap Between Adversarial ML Research and Practice
Giovanni Apruzzese
Hyrum S. Anderson
Savino Dambra
D. Freeman
Fabio Pierazzi
Kevin A. Roundy
AAML
14
74
0
29 Dec 2022
When Federated Learning Meets Pre-trained Language Models' Parameter-Efficient Tuning Methods
Zhuo Zhang
Yuanhang Yang
Yong Dai
Lizhen Qu
Zenglin Xu
FedML
8
63
0
20 Dec 2022
Why So Toxic? Measuring and Triggering Toxic Behavior in Open-Domain Chatbots
Waiman Si
Michael Backes
Jeremy Blackburn
Emiliano De Cristofaro
Gianluca Stringhini
Savvas Zannettou
Yang Zhang
11
57
0
07 Sep 2022
Differential Privacy in Natural Language Processing: The Story So Far
Oleksandra Klymenko
Stephen Meisenbacher
Florian Matthes
19
15
0
17 Aug 2022
Recovering Private Text in Federated Learning of Language Models
Samyak Gupta
Yangsibo Huang
Zexuan Zhong
Tianyu Gao
Kai Li
Danqi Chen
FedML
17
74
0
17 May 2022
You Are What You Write: Preserving Privacy in the Era of Large Language Models
Richard Plant
V. Giuffrida
Dimitra Gkatzia
PILM
8
19
0
20 Apr 2022
Quantifying Privacy Risks of Masked Language Models Using Membership Inference Attacks
Fatemehsadat Mireshghallah
Kartik Goyal
Archit Uniyal
Taylor Berg-Kirkpatrick
Reza Shokri
MIALM
18
151
0
08 Mar 2022
Membership Inference on Word Embedding and Beyond
Saeed Mahloujifar
Huseyin A. Inan
Melissa Chase
Esha Ghosh
Marcello Hasegawa
MIACV
SILM
8
46
0
21 Jun 2021
Membership Inference Attacks on Machine Learning: A Survey
Hongsheng Hu
Z. Salcic
Lichao Sun
Gillian Dobbie
Philip S. Yu
Xuyun Zhang
MIACV
11
408
0
14 Mar 2021
Quantifying and Mitigating Privacy Risks of Contrastive Learning
Xinlei He
Yang Zhang
11
50
0
08 Feb 2021
A Survey of Privacy Attacks in Machine Learning
M. Rigaki
Sebastian Garcia
PILM
AAML
6
210
0
15 Jul 2020
Convolutional Neural Networks for Sentence Classification
Yoon Kim
AILaw
VLM
241
13,283
0
25 Aug 2014
1