ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1802.08232
  4. Cited By
The Secret Sharer: Evaluating and Testing Unintended Memorization in
  Neural Networks

The Secret Sharer: Evaluating and Testing Unintended Memorization in Neural Networks

22 February 2018
Nicholas Carlini
Chang-rui Liu
Ulfar Erlingsson
Jernej Kos
D. Song
ArXivPDFHTML

Papers citing "The Secret Sharer: Evaluating and Testing Unintended Memorization in Neural Networks"

50 / 710 papers shown
Title
Demystifying Verbatim Memorization in Large Language Models
Demystifying Verbatim Memorization in Large Language Models
Jing Huang
Diyi Yang
Christopher Potts
ELM
PILM
MU
47
19
0
25 Jul 2024
Reconstructing Training Data From Real World Models Trained with
  Transfer Learning
Reconstructing Training Data From Real World Models Trained with Transfer Learning
Yakir Oz
Gilad Yehudai
Gal Vardi
Itai Antebi
Michal Irani
Niv Haim
32
2
0
22 Jul 2024
Weights Shuffling for Improving DPSGD in Transformer-based Models
Weights Shuffling for Improving DPSGD in Transformer-based Models
Jungang Yang
Zhe Ji
Liyao Xiang
35
0
0
22 Jul 2024
Operationalizing a Threat Model for Red-Teaming Large Language Models
  (LLMs)
Operationalizing a Threat Model for Red-Teaming Large Language Models (LLMs)
Apurv Verma
Satyapriya Krishna
Sebastian Gehrmann
Madhavan Seshadri
Anu Pradhan
Tom Ault
Leslie Barrett
David Rabinowitz
John Doucette
Nhathai Phan
51
9
0
20 Jul 2024
Feature Inference Attack on Shapley Values
Feature Inference Attack on Shapley Values
Xinjian Luo
Yangfan Jiang
X. Xiao
AAML
FAtt
30
19
0
16 Jul 2024
Social and Ethical Risks Posed by General-Purpose LLMs for Settling
  Newcomers in Canada
Social and Ethical Risks Posed by General-Purpose LLMs for Settling Newcomers in Canada
I. Nejadgholi
Maryam Molamohammadi
Samir Bakhtawar
43
0
0
15 Jul 2024
Privacy-Preserving Collaborative Genomic Research: A Real-Life
  Deployment and Vision
Privacy-Preserving Collaborative Genomic Research: A Real-Life Deployment and Vision
Zahra Rahmani
Nahal Shahini
Nadav Gat
Zebin Yun
Yuzhou Jiang
Ofir Farchy
Yaniv Harel
Vipin Chaudhary
Mahmood Sharif
Erman Ayday
SyDa
46
1
0
12 Jul 2024
Extracting Training Data from Document-Based VQA Models
Extracting Training Data from Document-Based VQA Models
Francesco Pinto
N. Rauschmayr
F. Tramèr
Philip H. S. Torr
Federico Tombari
29
3
0
11 Jul 2024
Fine-Tuning Large Language Models with User-Level Differential Privacy
Fine-Tuning Large Language Models with User-Level Differential Privacy
Zachary Charles
Arun Ganesh
Ryan McKenna
H. B. McMahan
Nicole Mitchell
Krishna Pillutla
Keith Rush
33
11
0
10 Jul 2024
Composable Interventions for Language Models
Composable Interventions for Language Models
Arinbjorn Kolbeinsson
Kyle O'Brien
Tianjin Huang
Shanghua Gao
Shiwei Liu
...
Anurag J. Vaidya
Faisal Mahmood
Marinka Zitnik
Tianlong Chen
Thomas Hartvigsen
KELM
MU
87
5
0
09 Jul 2024
Exposing Privacy Gaps: Membership Inference Attack on Preference Data for LLM Alignment
Exposing Privacy Gaps: Membership Inference Attack on Preference Data for LLM Alignment
Qizhang Feng
Siva Rajesh Kasa
Santhosh Kumar Kasa
Hyokun Yun
C. Teo
S. Bodapati
84
6
0
08 Jul 2024
Releasing Malevolence from Benevolence: The Menace of Benign Data on
  Machine Unlearning
Releasing Malevolence from Benevolence: The Menace of Benign Data on Machine Unlearning
Binhao Ma
Tianhang Zheng
Hongsheng Hu
Di Wang
Shuo Wang
Zhongjie Ba
Zhan Qin
Kui Ren
AAML
28
3
0
06 Jul 2024
Towards More Realistic Extraction Attacks: An Adversarial Perspective
Towards More Realistic Extraction Attacks: An Adversarial Perspective
Yash More
Prakhar Ganesh
G. Farnadi
AAML
66
6
0
02 Jul 2024
The Art of Saying No: Contextual Noncompliance in Language Models
The Art of Saying No: Contextual Noncompliance in Language Models
Faeze Brahman
Sachin Kumar
Vidhisha Balachandran
Pradeep Dasigi
Valentina Pyatkin
...
Jack Hessel
Yulia Tsvetkov
Noah A. Smith
Yejin Choi
Hannaneh Hajishirzi
69
20
0
02 Jul 2024
A Study of Nationality Bias in Names and Perplexity using Off-the-Shelf
  Affect-related Tweet Classifiers
A Study of Nationality Bias in Names and Perplexity using Off-the-Shelf Affect-related Tweet Classifiers
Valentin Barriere
Sebastian Cifuentes
28
0
0
01 Jul 2024
Silver Linings in the Shadows: Harnessing Membership Inference for
  Machine Unlearning
Silver Linings in the Shadows: Harnessing Membership Inference for Machine Unlearning
Nexhi Sula
Abhinav Kumar
Jie Hou
Han Wang
R. Tourani
MU
23
0
0
01 Jul 2024
Characterizing Stereotypical Bias from Privacy-preserving Pre-Training
Characterizing Stereotypical Bias from Privacy-preserving Pre-Training
Stefan Arnold
Rene Gröbner
Annika Schreiner
37
0
0
30 Jun 2024
LongLaMP: A Benchmark for Personalized Long-form Text Generation
LongLaMP: A Benchmark for Personalized Long-form Text Generation
Ishita Kumar
Snigdha Viswanathan
Sushrita Yerra
Alireza Salemi
Ryan A. Rossi
...
Xiang Chen
Ruiyi Zhang
Shubham Agarwal
Nedim Lipka
Hamed Zamani
41
22
0
27 Jun 2024
Evaluating Copyright Takedown Methods for Language Models
Evaluating Copyright Takedown Methods for Language Models
Boyi Wei
Weijia Shi
Yangsibo Huang
Noah A. Smith
Chiyuan Zhang
Luke Zettlemoyer
Kai Li
Peter Henderson
49
19
0
26 Jun 2024
Enhancing Federated Learning with Adaptive Differential Privacy and
  Priority-Based Aggregation
Enhancing Federated Learning with Adaptive Differential Privacy and Priority-Based Aggregation
Mahtab Talaei
Iman Izadi
FedML
24
0
0
26 Jun 2024
Enhancing Data Privacy in Large Language Models through Private
  Association Editing
Enhancing Data Privacy in Large Language Models through Private Association Editing
Davide Venditti
Elena Sofia Ruzzetti
Giancarlo A. Xompero
Cristina Giannone
Andrea Favalli
Raniero Romagnoli
Fabio Massimo Zanzotto
KELM
35
2
0
26 Jun 2024
Recite, Reconstruct, Recollect: Memorization in LMs as a Multifaceted Phenomenon
Recite, Reconstruct, Recollect: Memorization in LMs as a Multifaceted Phenomenon
USVSN Sai Prashanth
Alvin Deng
Kyle O'Brien
Jyothir S V
Mohammad Aflah Khan
...
Jacob Ray Fuehne
Stella Biderman
Tracy Ke
Katherine Lee
Naomi Saphra
55
12
0
25 Jun 2024
Noisy Neighbors: Efficient membership inference attacks against LLMs
Noisy Neighbors: Efficient membership inference attacks against LLMs
Filippo Galli
Luca Melis
Tommaso Cucinotta
44
7
0
24 Jun 2024
A Tale of Trust and Accuracy: Base vs. Instruct LLMs in RAG Systems
A Tale of Trust and Accuracy: Base vs. Instruct LLMs in RAG Systems
Florin Cuconasu
Giovanni Trappolini
Nicola Tonellotto
Fabrizio Silvestri
51
2
0
21 Jun 2024
Unmasking Database Vulnerabilities: Zero-Knowledge Schema Inference
  Attacks in Text-to-SQL Systems
Unmasking Database Vulnerabilities: Zero-Knowledge Schema Inference Attacks in Text-to-SQL Systems
Đorđe Klisura
Anthony Rios
AAML
24
1
0
20 Jun 2024
Protecting Privacy Through Approximating Optimal Parameters for Sequence
  Unlearning in Language Models
Protecting Privacy Through Approximating Optimal Parameters for Sequence Unlearning in Language Models
Dohyun Lee
Daniel Rim
Minseok Choi
Jaegul Choo
PILM
MU
57
4
0
20 Jun 2024
AspirinSum: an Aspect-based utility-preserved de-identification
  Summarization framework
AspirinSum: an Aspect-based utility-preserved de-identification Summarization framework
Ya-Lun Li
45
0
0
20 Jun 2024
Be like a Goldfish, Don't Memorize! Mitigating Memorization in
  Generative LLMs
Be like a Goldfish, Don't Memorize! Mitigating Memorization in Generative LLMs
Abhimanyu Hans
Yuxin Wen
Neel Jain
John Kirchenbauer
Hamid Kazemi
...
Siddharth Singh
Gowthami Somepalli
Jonas Geiping
A. Bhatele
Tom Goldstein
33
30
0
14 Jun 2024
REVS: Unlearning Sensitive Information in Language Models via Rank Editing in the Vocabulary Space
REVS: Unlearning Sensitive Information in Language Models via Rank Editing in the Vocabulary Space
Tomer Ashuach
Martin Tutek
Yonatan Belinkov
KELM
MU
65
4
0
13 Jun 2024
On the Minimal Degree Bias in Generalization on the Unseen for
  non-Boolean Functions
On the Minimal Degree Bias in Generalization on the Unseen for non-Boolean Functions
Denys Pushkin
Raphael Berthier
Emmanuel Abbe
30
0
0
10 Jun 2024
A Survey on Machine Unlearning: Techniques and New Emerged Privacy Risks
A Survey on Machine Unlearning: Techniques and New Emerged Privacy Risks
Hengzhu Liu
Ping Xiong
Tianqing Zhu
Philip S. Yu
27
6
0
10 Jun 2024
Causal Estimation of Memorisation Profiles
Causal Estimation of Memorisation Profiles
Pietro Lesci
Clara Meister
Thomas Hofmann
Andreas Vlachos
Tiago Pimentel
45
5
0
06 Jun 2024
Memorization in deep learning: A survey
Memorization in deep learning: A survey
Jiaheng Wei
Yanjun Zhang
Leo Yu Zhang
Ming Ding
Chao Chen
Kok-Leong Ong
Jun Zhang
Yang Xiang
42
6
0
06 Jun 2024
Finding NeMo: Localizing Neurons Responsible For Memorization in
  Diffusion Models
Finding NeMo: Localizing Neurons Responsible For Memorization in Diffusion Models
Dominik Hintersdorf
Lukas Struppek
Kristian Kersting
Adam Dziedzic
Franziska Boenisch
31
11
0
04 Jun 2024
Inference Attacks: A Taxonomy, Survey, and Promising Directions
Inference Attacks: A Taxonomy, Survey, and Promising Directions
Feng Wu
Lei Cui
Shaowen Yao
Shui Yu
39
2
0
04 Jun 2024
A Novel Review of Stability Techniques for Improved Privacy-Preserving
  Machine Learning
A Novel Review of Stability Techniques for Improved Privacy-Preserving Machine Learning
Coleman DuPlessie
Aidan Gao
23
0
0
31 May 2024
AI Risk Management Should Incorporate Both Safety and Security
AI Risk Management Should Incorporate Both Safety and Security
Xiangyu Qi
Yangsibo Huang
Yi Zeng
Edoardo Debenedetti
Jonas Geiping
...
Chaowei Xiao
Bo-wen Li
Dawn Song
Peter Henderson
Prateek Mittal
AAML
48
10
0
29 May 2024
Delving into Differentially Private Transformer
Delving into Differentially Private Transformer
Youlong Ding
Xueyang Wu
Yining Meng
Yonggang Luo
Hao Wang
Weike Pan
29
5
0
28 May 2024
Cross-Modal Safety Alignment: Is textual unlearning all you need?
Cross-Modal Safety Alignment: Is textual unlearning all you need?
Trishna Chakraborty
Erfan Shayegani
Zikui Cai
Nael B. Abu-Ghazaleh
M. Salman Asif
Yue Dong
A. Roy-Chowdhury
Chengyu Song
39
15
0
27 May 2024
OSLO: One-Shot Label-Only Membership Inference Attacks
OSLO: One-Shot Label-Only Membership Inference Attacks
Yuefeng Peng
Jaechul Roh
Subhransu Maji
Amir Houmansadr
44
0
0
27 May 2024
Data Reconstruction: When You See It and When You Don't
Data Reconstruction: When You See It and When You Don't
Edith Cohen
Haim Kaplan
Yishay Mansour
Shay Moran
Kobbi Nissim
Uri Stemmer
Eliad Tsfadia
AAML
42
2
0
24 May 2024
Better Membership Inference Privacy Measurement through Discrepancy
Better Membership Inference Privacy Measurement through Discrepancy
Ruihan Wu
Pengrun Huang
Kamalika Chaudhuri
MIACV
32
0
0
24 May 2024
The Mosaic Memory of Large Language Models
The Mosaic Memory of Large Language Models
Igor Shilov
Matthieu Meeus
Yves-Alexandre de Montjoye
39
3
0
24 May 2024
Tiny Refinements Elicit Resilience: Toward Efficient Prefix-Model
  Against LLM Red-Teaming
Tiny Refinements Elicit Resilience: Toward Efficient Prefix-Model Against LLM Red-Teaming
Jiaxu Liu
Xiangyu Yin
Sihao Wu
Jianhong Wang
Meng Fang
Xinping Yi
Xiaowei Huang
32
4
0
21 May 2024
Data Contamination Calibration for Black-box LLMs
Data Contamination Calibration for Black-box LLMs
Wen-song Ye
Jiaqi Hu
Liyao Li
Haobo Wang
Gang Chen
Junbo Zhao
34
6
0
20 May 2024
Quantifying In-Context Reasoning Effects and Memorization Effects in
  LLMs
Quantifying In-Context Reasoning Effects and Memorization Effects in LLMs
Siyu Lou
Yuntian Chen
Xiaodan Liang
Liang Lin
Quanshi Zhang
32
2
0
20 May 2024
SecureLLM: Using Compositionality to Build Provably Secure Language
  Models for Private, Sensitive, and Secret Data
SecureLLM: Using Compositionality to Build Provably Secure Language Models for Private, Sensitive, and Secret Data
Abdulrahman Alabdulakreem
Christian M Arnold
Yerim Lee
Pieter M Feenstra
Boris Katz
Andrei Barbu
16
4
0
16 May 2024
GLiRA: Black-Box Membership Inference Attack via Knowledge Distillation
GLiRA: Black-Box Membership Inference Attack via Knowledge Distillation
Andrey V. Galichin
Mikhail Aleksandrovich Pautov
Alexey Zhavoronkin
Oleg Y. Rogov
Ivan V. Oseledets
AAML
21
1
0
13 May 2024
Shadow-Free Membership Inference Attacks: Recommender Systems Are More
  Vulnerable Than You Thought
Shadow-Free Membership Inference Attacks: Recommender Systems Are More Vulnerable Than You Thought
Xiaoxiao Chi
Xuyun Zhang
Yan Wang
Lianyong Qi
Amin Beheshti
Xiaolong Xu
Kim-Kwang Raymond Choo
Shuo Wang
Hongsheng Hu
28
0
0
11 May 2024
Special Characters Attack: Toward Scalable Training Data Extraction From
  Large Language Models
Special Characters Attack: Toward Scalable Training Data Extraction From Large Language Models
Yang Bai
Ge Pei
Jindong Gu
Yong Yang
Xingjun Ma
31
10
0
09 May 2024
Previous
123456...131415
Next