Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
1807.05185
Cited By
Model Reconstruction from Model Explanations
13 July 2018
S. Milli
Ludwig Schmidt
Anca Dragan
Moritz Hardt
FAtt
Re-assign community
ArXiv
PDF
HTML
Papers citing
"Model Reconstruction from Model Explanations"
36 / 36 papers shown
Title
Crowding Out The Noise: Algorithmic Collective Action Under Differential Privacy
Rushabh Solanki
Meghana Bhange
Ulrich Aïvodji
Elliot Creager
29
0
0
09 May 2025
Privacy Risks and Preservation Methods in Explainable Artificial Intelligence: A Scoping Review
Sonal Allana
Mohan Kankanhalli
Rozita Dara
32
0
0
05 May 2025
Attackers Can Do Better: Over- and Understated Factors of Model Stealing Attacks
Daryna Oliynyk
Rudolf Mayer
Andreas Rauber
AAML
46
0
0
08 Mar 2025
Examining the Threat Landscape: Foundation Models and Model Stealing
Ankita Raj
Deepankar Varma
Chetan Arora
AAML
68
1
0
25 Feb 2025
Polynomial Time Cryptanalytic Extraction of Deep Neural Networks in the Hard-Label Setting
Nicholas Carlini
J. Chávez-Saab
Anna Hambitzer
Francisco Rodríguez-Henríquez
Adi Shamir
AAML
27
1
0
08 Oct 2024
Feature Inference Attack on Shapley Values
Xinjian Luo
Yangfan Jiang
X. Xiao
AAML
FAtt
30
19
0
16 Jul 2024
Isolation and Induction: Training Robust Deep Neural Networks against Model Stealing Attacks
Jun Guo
Aishan Liu
Xingyu Zheng
Siyuan Liang
Yisong Xiao
Yichao Wu
Xianglong Liu
AAML
30
12
0
02 Aug 2023
The Case Against Explainability
Hofit Wasserman Rozen
N. Elkin-Koren
Ran Gilad-Bachrach
AILaw
ELM
21
1
0
20 May 2023
Finding Meaningful Distributions of ML Black-boxes under Forensic Investigation
Jiyi Zhang
Hansheng Fang
Hwee Kuan Lee
E. Chang
16
1
0
10 May 2023
Model Extraction Attacks on Split Federated Learning
Jingtao Li
Adnan Siraj Rakin
Xing Chen
Li Yang
Zhezhi He
Deliang Fan
C. Chakrabarti
FedML
55
5
0
13 Mar 2023
AUTOLYCUS: Exploiting Explainable AI (XAI) for Model Extraction Attacks against Interpretable Models
Abdullah Çaglar Öksüz
Anisa Halimi
Erman Ayday
ELM
AAML
21
2
0
04 Feb 2023
Tensions Between the Proxies of Human Values in AI
Teresa Datta
D. Nissani
Max Cembalest
Akash Khanna
Haley Massa
John P. Dickerson
28
2
0
14 Dec 2022
Differentially Private Counterfactuals via Functional Mechanism
Fan Yang
Qizhang Feng
Kaixiong Zhou
Jiahao Chen
Xia Hu
24
8
0
04 Aug 2022
I Know What You Trained Last Summer: A Survey on Stealing Machine Learning Models and Defences
Daryna Oliynyk
Rudolf Mayer
Andreas Rauber
39
106
0
16 Jun 2022
Reconstructing Training Data from Trained Neural Networks
Niv Haim
Gal Vardi
Gilad Yehudai
Ohad Shamir
Michal Irani
40
132
0
15 Jun 2022
Margin-distancing for safe model explanation
Tom Yan
Chicheng Zhang
28
3
0
23 Feb 2022
Hardness of Noise-Free Learning for Two-Hidden-Layer Neural Networks
Sitan Chen
Aravind Gollakota
Adam R. Klivans
Raghu Meka
21
30
0
10 Feb 2022
Mitigating Adversarial Attacks by Distributing Different Copies to Different Users
Jiyi Zhang
Hansheng Fang
W. Tann
Ke Xu
Chengfang Fang
E. Chang
AAML
21
3
0
30 Nov 2021
Efficiently Learning Any One Hidden Layer ReLU Network From Queries
Sitan Chen
Adam R. Klivans
Raghu Meka
MLAU
MLT
42
8
0
08 Nov 2021
DeepSteal: Advanced Model Extractions Leveraging Efficient Weight Stealing in Memories
Adnan Siraj Rakin
Md Hafizul Islam Chowdhuryy
Fan Yao
Deliang Fan
AAML
MIACV
36
110
0
08 Nov 2021
SoK: Machine Learning Governance
Varun Chandrasekaran
Hengrui Jia
Anvith Thudi
Adelin Travers
Mohammad Yaghini
Nicolas Papernot
30
16
0
20 Sep 2021
Amazon SageMaker Clarify: Machine Learning Bias Detection and Explainability in the Cloud
Michaela Hardt
Xiaoguang Chen
Xiaoyi Cheng
Michele Donini
J. Gelman
...
Muhammad Bilal Zafar
Sanjiv Ranjan Das
Kevin Haas
Tyler Hill
K. Kenthapadi
ELM
FaML
25
42
0
07 Sep 2021
MEGEX: Data-Free Model Extraction Attack against Gradient-Based Explainable AI
T. Miura
Satoshi Hasegawa
Toshiki Shibahara
SILM
MIACV
16
37
0
19 Jul 2021
Survey: Leakage and Privacy at Inference Time
Marija Jegorova
Chaitanya Kaul
Charlie Mayor
Alison Q. OÑeil
Alexander Weir
Roderick Murray-Smith
Sotirios A. Tsaftaris
PILM
MIACV
17
71
0
04 Jul 2021
Characterizing the risk of fairwashing
Ulrich Aivodji
Hiromi Arai
Sébastien Gambs
Satoshi Hara
20
27
0
14 Jun 2021
Exploiting Explanations for Model Inversion Attacks
Xu Zhao
Wencan Zhang
Xiao Xiao
Brian Y. Lim
MIACV
21
82
0
26 Apr 2021
Explanation from Specification
Harish Naik
Gyorgy Turán
XAI
21
0
0
13 Dec 2020
Data-Free Model Extraction
Jean-Baptiste Truong
Pratyush Maini
R. Walls
Nicolas Papernot
MIACV
15
181
0
30 Nov 2020
Model extraction from counterfactual explanations
Ulrich Aivodji
Alexandre Bolot
Sébastien Gambs
MIACV
MLAU
27
51
0
03 Sep 2020
Simulating Unknown Target Models for Query-Efficient Black-box Attacks
Chen Ma
L. Chen
Junhai Yong
MLAU
OOD
39
17
0
02 Sep 2020
A Survey of Privacy Attacks in Machine Learning
M. Rigaki
Sebastian Garcia
PILM
AAML
27
213
0
15 Jul 2020
The Trade-Offs of Private Prediction
L. V. D. van der Maaten
Awni Y. Hannun
15
22
0
09 Jul 2020
Model Explanations with Differential Privacy
Neel Patel
Reza Shokri
Yair Zick
SILM
FedML
13
32
0
16 Jun 2020
Stealing Deep Reinforcement Learning Models for Fun and Profit
Kangjie Chen
Shangwei Guo
Tianwei Zhang
Xiaofei Xie
Yang Liu
MLAU
MIACV
OffRL
14
45
0
09 Jun 2020
SNIFF: Reverse Engineering of Neural Networks with Fault Attacks
J. Breier
Dirmanto Jap
Xiaolu Hou
S. Bhasin
Yang Liu
15
52
0
23 Feb 2020
Mind Your Weight(s): A Large-scale Study on Insufficient Machine Learning Model Protection in Mobile Apps
Zhichuang Sun
Ruimin Sun
Long Lu
Alan Mislove
28
78
0
18 Feb 2020
1