ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1901.02402
  4. Cited By
Contamination Attacks and Mitigation in Multi-Party Machine Learning

Contamination Attacks and Mitigation in Multi-Party Machine Learning

8 January 2019
Jamie Hayes
O. Ohrimenko
    AAML
    FedML
ArXivPDFHTML

Papers citing "Contamination Attacks and Mitigation in Multi-Party Machine Learning"

25 / 25 papers shown
Title
Support Vector Machines under Adversarial Label Contamination
Support Vector Machines under Adversarial Label Contamination
Huang Xiao
Battista Biggio
B. Nelson
Han Xiao
Claudia Eckert
Fabio Roli
AAML
31
231
0
01 Jun 2022
Machine Learning with Membership Privacy using Adversarial
  Regularization
Machine Learning with Membership Privacy using Adversarial Regularization
Milad Nasr
Reza Shokri
Amir Houmansadr
FedML
MIACV
35
468
0
16 Jul 2018
An Algorithmic Framework For Differentially Private Data Analysis on
  Trusted Processors
An Algorithmic Framework For Differentially Private Data Analysis on Trusted Processors
Joshua Allen
Bolin Ding
Janardhan Kulkarni
Harsha Nori
O. Ohrimenko
Sergey Yekhanin
SyDa
FedML
89
32
0
02 Jul 2018
Is feature selection secure against training data poisoning?
Is feature selection secure against training data poisoning?
Huang Xiao
Battista Biggio
Gavin Brown
Giorgio Fumera
Claudia Eckert
Fabio Roli
AAML
SILM
39
423
0
21 Apr 2018
Manipulating Machine Learning: Poisoning Attacks and Countermeasures for
  Regression Learning
Manipulating Machine Learning: Poisoning Attacks and Countermeasures for Regression Learning
Matthew Jagielski
Alina Oprea
Battista Biggio
Chang-rui Liu
Cristina Nita-Rotaru
Yue Liu
AAML
57
757
0
01 Apr 2018
Understanding Membership Inferences on Well-Generalized Learning Models
Understanding Membership Inferences on Well-Generalized Learning Models
Yunhui Long
Vincent Bindschaedler
Lei Wang
Diyue Bu
Xiaofeng Wang
Haixu Tang
Carl A. Gunter
Kai Chen
MIALM
MIACV
29
224
0
13 Feb 2018
Targeted Backdoor Attacks on Deep Learning Systems Using Data Poisoning
Targeted Backdoor Attacks on Deep Learning Systems Using Data Poisoning
Xinyun Chen
Chang-rui Liu
Yue Liu
Kimberly Lu
D. Song
AAML
SILM
78
1,822
0
15 Dec 2017
Prochlo: Strong Privacy for Analytics in the Crowd
Prochlo: Strong Privacy for Analytics in the Crowd
Andrea Bittau
Ulfar Erlingsson
Petros Maniatis
Ilya Mironov
A. Raghunathan
David Lie
Mitch Rudominer
Ushasree Kode
J. Tinnés
B. Seefeld
106
279
0
02 Oct 2017
Machine Learning Models that Remember Too Much
Machine Learning Models that Remember Too Much
Congzheng Song
Thomas Ristenpart
Vitaly Shmatikov
VLM
48
508
0
22 Sep 2017
BadNets: Identifying Vulnerabilities in the Machine Learning Model
  Supply Chain
BadNets: Identifying Vulnerabilities in the Machine Learning Model Supply Chain
Tianyu Gu
Brendan Dolan-Gavitt
S. Garg
SILM
72
1,754
0
22 Aug 2017
Understanding Black-box Predictions via Influence Functions
Understanding Black-box Predictions via Influence Functions
Pang Wei Koh
Percy Liang
TDI
134
2,854
0
14 Mar 2017
Deep Models Under the GAN: Information Leakage from Collaborative Deep
  Learning
Deep Models Under the GAN: Information Leakage from Collaborative Deep Learning
Briland Hitaj
G. Ateniese
Fernando Perez-Cruz
FedML
105
1,385
0
24 Feb 2017
Learning to Pivot with Adversarial Networks
Learning to Pivot with Adversarial Networks
Gilles Louppe
Michael Kagan
Kyle Cranmer
46
227
0
03 Nov 2016
Membership Inference Attacks against Machine Learning Models
Membership Inference Attacks against Machine Learning Models
Reza Shokri
M. Stronati
Congzheng Song
Vitaly Shmatikov
SLR
MIALM
MIACV
200
4,075
0
18 Oct 2016
Minimax Filter: Learning to Preserve Privacy from Inference Attacks
Minimax Filter: Learning to Preserve Privacy from Inference Attacks
Jihun Hamm
23
82
0
12 Oct 2016
Towards Evaluating the Robustness of Neural Networks
Towards Evaluating the Robustness of Neural Networks
Nicholas Carlini
D. Wagner
OOD
AAML
160
8,497
0
16 Aug 2016
Adversarial examples in the physical world
Adversarial examples in the physical world
Alexey Kurakin
Ian Goodfellow
Samy Bengio
SILM
AAML
489
5,868
0
08 Jul 2016
Deep Learning with Differential Privacy
Deep Learning with Differential Privacy
Martín Abadi
Andy Chu
Ian Goodfellow
H. B. McMahan
Ilya Mironov
Kunal Talwar
Li Zhang
FedML
SyDa
162
6,049
0
01 Jul 2016
Learning Privately from Multiparty Data
Learning Privately from Multiparty Data
Jihun Hamm
Yingjun Cao
M. Belkin
FedML
31
165
0
10 Feb 2016
The Limitations of Deep Learning in Adversarial Settings
The Limitations of Deep Learning in Adversarial Settings
Nicolas Papernot
Patrick McDaniel
S. Jha
Matt Fredrikson
Z. Berkay Celik
A. Swami
AAML
60
3,947
0
24 Nov 2015
Censoring Representations with an Adversary
Censoring Representations with an Adversary
Harrison Edwards
Amos Storkey
AAML
FaML
47
504
0
18 Nov 2015
DeepFool: a simple and accurate method to fool deep neural networks
DeepFool: a simple and accurate method to fool deep neural networks
Seyed-Mohsen Moosavi-Dezfooli
Alhussein Fawzi
P. Frossard
AAML
90
4,878
0
14 Nov 2015
Explaining and Harnessing Adversarial Examples
Explaining and Harnessing Adversarial Examples
Ian Goodfellow
Jonathon Shlens
Christian Szegedy
AAML
GAN
158
18,922
0
20 Dec 2014
Convolutional Neural Networks for Sentence Classification
Convolutional Neural Networks for Sentence Classification
Yoon Kim
AILaw
VLM
541
13,395
0
25 Aug 2014
Poisoning Attacks against Support Vector Machines
Poisoning Attacks against Support Vector Machines
Battista Biggio
B. Nelson
Pavel Laskov
AAML
80
1,580
0
27 Jun 2012
1