ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1911.07399
  4. Cited By
NeuronInspect: Detecting Backdoors in Neural Networks via Output
  Explanations

NeuronInspect: Detecting Backdoors in Neural Networks via Output Explanations

18 November 2019
Xijie Huang
M. Alzantot
Mani B. Srivastava
    AAML
ArXiv (abs)PDFHTML

Papers citing "NeuronInspect: Detecting Backdoors in Neural Networks via Output Explanations"

50 / 58 papers shown
Enhancing the Effectiveness and Durability of Backdoor Attacks in Federated Learning through Maximizing Task Distinction
Enhancing the Effectiveness and Durability of Backdoor Attacks in Federated Learning through Maximizing Task Distinction
Zhaoxin Wang
Handing Wang
Cong Tian
Yaochu Jin
FedMLAAML
144
0
0
23 Sep 2025
BackdoorMBTI: A Backdoor Learning Multimodal Benchmark Tool Kit for Backdoor Defense EvaluationKnowledge Discovery and Data Mining (KDD), 2024
Haiyang Yu
Tian Xie
Jiaping Gui
Pengyang Wang
P. Yi
Yue Wu
355
2
0
17 Nov 2024
A Practical Trigger-Free Backdoor Attack on Neural Networks
A Practical Trigger-Free Backdoor Attack on Neural Networks
Jiahao Wang
Xianglong Zhang
Xiuzhen Cheng
Pengfei Hu
Guoming Zhang
AAML
191
1
0
21 Aug 2024
A Survey of Trojan Attacks and Defenses to Deep Neural Networks
A Survey of Trojan Attacks and Defenses to Deep Neural Networks
Lingxin Jin
Xianyu Wen
Wei Jiang
Jinyu Zhan
AAML
230
3
0
15 Aug 2024
Clean-Label Physical Backdoor Attacks with Data Distillation
Clean-Label Physical Backdoor Attacks with Data Distillation
Thinh Dao
Cuong Chi Le
Khoa D. Doan
AAML
481
1
0
27 Jul 2024
Watch the Watcher! Backdoor Attacks on Security-Enhancing Diffusion
  Models
Watch the Watcher! Backdoor Attacks on Security-Enhancing Diffusion Models
Changjiang Li
Ren Pang
Bochuan Cao
Jinghui Chen
Fenglong Ma
Shouling Ji
Ting Wang
DiffM
181
4
0
14 Jun 2024
Generalization Bound and New Algorithm for Clean-Label Backdoor Attack
Generalization Bound and New Algorithm for Clean-Label Backdoor Attack
Lijia Yu
Shuang Liu
Yibo Miao
Xiao-Shan Gao
Lijun Zhang
AAML
304
10
0
02 Jun 2024
BackdoorIndicator: Leveraging OOD Data for Proactive Backdoor Detection
  in Federated Learning
BackdoorIndicator: Leveraging OOD Data for Proactive Backdoor Detection in Federated Learning
Songze Li
Yanbo Dai
AAMLFedML
228
20
0
31 May 2024
Not All Prompts Are Secure: A Switchable Backdoor Attack Against
  Pre-trained Vision Transformers
Not All Prompts Are Secure: A Switchable Backdoor Attack Against Pre-trained Vision Transformers
Shengyuan Yang
Jiawang Bai
Kuofeng Gao
Yong-Liang Yang
Yiming Li
Shu-Tao Xia
AAMLSILM
315
5
0
17 May 2024
A Backdoor-based Explainable AI Benchmark for High Fidelity Evaluation of Attributions
A Backdoor-based Explainable AI Benchmark for High Fidelity Evaluation of Attributions
Peiyu Yang
Naveed Akhtar
Jiantong Jiang
Lin Wang
XAI
225
2
0
02 May 2024
LOTUS: Evasive and Resilient Backdoor Attacks through Sub-Partitioning
LOTUS: Evasive and Resilient Backdoor Attacks through Sub-Partitioning
Shuyang Cheng
Guanhong Tao
Yingqi Liu
Guangyu Shen
Shengwei An
Shiwei Feng
Xiangzhe Xu
Kaiyuan Zhang
Shiqing Ma
Xiangyu Zhang
AAML
227
10
0
25 Mar 2024
Game of Trojans: Adaptive Adversaries Against Output-based
  Trojaned-Model Detectors
Game of Trojans: Adaptive Adversaries Against Output-based Trojaned-Model Detectors
D. Sahabandu
Xiaojun Xu
Arezoo Rajabi
Luyao Niu
Bhaskar Ramasubramanian
Bo Li
Radha Poovendran
AAML
188
1
0
12 Feb 2024
Preference Poisoning Attacks on Reward Model Learning
Preference Poisoning Attacks on Reward Model Learning
Junlin Wu
Zhenghao Hu
Chaowei Xiao
Chenguang Wang
Ning Zhang
Yevgeniy Vorobeychik
AAML
276
11
0
02 Feb 2024
UltraClean: A Simple Framework to Train Robust Neural Networks against Backdoor Attacks
UltraClean: A Simple Framework to Train Robust Neural Networks against Backdoor Attacks
Bingyin Zhao
Yingjie Lao
AAML
300
2
0
17 Dec 2023
On the Difficulty of Defending Contrastive Learning against Backdoor
  Attacks
On the Difficulty of Defending Contrastive Learning against Backdoor AttacksUSENIX Security Symposium (USENIX Security), 2023
Changjiang Li
Ren Pang
Bochuan Cao
Zhaohan Xi
Jinghui Chen
R. Beyah
Ting Wang
AAML
286
7
0
14 Dec 2023
Towards Sample-specific Backdoor Attack with Clean Labels via Attribute Trigger
Towards Sample-specific Backdoor Attack with Clean Labels via Attribute TriggerIEEE Transactions on Dependable and Secure Computing (IEEE TDSC), 2023
Yiming Li
Mingyan Zhu
Junfeng Guo
Tao Wei
Shu-Tao Xia
Zhan Qin
AAML
373
5
0
03 Dec 2023
SecurityNet: Assessing Machine Learning Vulnerabilities on Public Models
SecurityNet: Assessing Machine Learning Vulnerabilities on Public Models
Boyang Zhang
Zheng Li
Ziqing Yang
Xinlei He
Michael Backes
Mario Fritz
Yang Zhang
337
8
0
19 Oct 2023
XGBD: Explanation-Guided Graph Backdoor Detection
XGBD: Explanation-Guided Graph Backdoor DetectionEuropean Conference on Artificial Intelligence (ECAI), 2023
Zihan Guan
Mengnan Du
Ninghao Liu
AAML
259
11
0
08 Aug 2023
TIJO: Trigger Inversion with Joint Optimization for Defending Multimodal
  Backdoored Models
TIJO: Trigger Inversion with Joint Optimization for Defending Multimodal Backdoored ModelsIEEE International Conference on Computer Vision (ICCV), 2023
Indranil Sur
Karan Sikka
Matthew Walmer
K. Koneripalli
Anirban Roy
Xiaoyu Lin
Ajay Divakaran
Susmit Jha
155
12
0
07 Aug 2023
A Survey of Safety and Trustworthiness of Large Language Models through
  the Lens of Verification and Validation
A Survey of Safety and Trustworthiness of Large Language Models through the Lens of Verification and ValidationArtificial Intelligence Review (AIR), 2023
Xiaowei Huang
Wenjie Ruan
Wei Huang
Gao Jin
Yizhen Dong
...
Sihao Wu
Peipei Xu
Dengyu Wu
André Freitas
Mustafa A. Mustafa
ALM
351
146
0
19 May 2023
Defending Against Patch-based Backdoor Attacks on Self-Supervised
  Learning
Defending Against Patch-based Backdoor Attacks on Self-Supervised LearningComputer Vision and Pattern Recognition (CVPR), 2023
Ajinkya Tejankar
Maziar Sanjabi
Qifan Wang
Sinong Wang
Hamed Firooz
Hamed Pirsiavash
L Tan
AAML
195
25
0
04 Apr 2023
Poisoning Web-Scale Training Datasets is Practical
Poisoning Web-Scale Training Datasets is PracticalIEEE Symposium on Security and Privacy (IEEE S&P), 2023
Nicholas Carlini
Matthew Jagielski
Christopher A. Choquette-Choo
Daniel Paleka
Will Pearce
Hyrum S. Anderson
Seth Neel
Kurt Thomas
Florian Tramèr
SILM
376
268
0
20 Feb 2023
Mithridates: Auditing and Boosting Backdoor Resistance of Machine
  Learning Pipelines
Mithridates: Auditing and Boosting Backdoor Resistance of Machine Learning PipelinesConference on Computer and Communications Security (CCS), 2023
Eugene Bagdasaryan
Vitaly Shmatikov
AAML
326
3
0
09 Feb 2023
BEAGLE: Forensics of Deep Learning Backdoor Attack for Better Defense
BEAGLE: Forensics of Deep Learning Backdoor Attack for Better DefenseNetwork and Distributed System Security Symposium (NDSS), 2023
Shuyang Cheng
Guanhong Tao
Yingqi Liu
Shengwei An
Xiangzhe Xu
...
Guangyu Shen
Kaiyuan Zhang
Qiuling Xu
Shiqing Ma
Xiangyu Zhang
AAML
214
20
0
16 Jan 2023
XMAM:X-raying Models with A Matrix to Reveal Backdoor Attacks for
  Federated Learning
XMAM:X-raying Models with A Matrix to Reveal Backdoor Attacks for Federated Learning
Jianyi Zhang
Fangjiao Zhang
Qichao Jin
Zhiqiang Wang
Xiaodong Lin
X. Hei
AAMLFedML
184
2
0
28 Dec 2022
Fine-Tuning Is All You Need to Mitigate Backdoor Attacks
Fine-Tuning Is All You Need to Mitigate Backdoor Attacks
Zeyang Sha
Xinlei He
Pascal Berrang
Mathias Humbert
Yang Zhang
AAML
262
49
0
18 Dec 2022
Be Careful with Rotation: A Uniform Backdoor Pattern for 3D Shape
Be Careful with Rotation: A Uniform Backdoor Pattern for 3D Shape
Linkun Fan
Fazhi He
Qingchen Guo
Wei Tang
Xiaolin Hong
Bing Li
AAML3DPC
337
9
0
28 Nov 2022
Dormant Neural Trojans
Dormant Neural TrojansInternational Conference on Machine Learning and Applications (ICMLA), 2022
Feisi Fu
Panagiota Kiourti
Wenchao Li
AAML
224
0
0
02 Nov 2022
An Embarrassingly Simple Backdoor Attack on Self-supervised Learning
An Embarrassingly Simple Backdoor Attack on Self-supervised LearningIEEE International Conference on Computer Vision (ICCV), 2022
Changjiang Li
Ren Pang
Zhaohan Xi
Tianyu Du
S. Ji
Yuan Yao
Ting Wang
AAML
319
41
0
13 Oct 2022
Understanding Impacts of Task Similarity on Backdoor Attack and
  Detection
Understanding Impacts of Task Similarity on Backdoor Attack and Detection
Di Tang
Rui Zhu
Luyi Xing
Haixu Tang
Yi Chen
AAML
262
5
0
12 Oct 2022
Untargeted Backdoor Watermark: Towards Harmless and Stealthy Dataset
  Copyright Protection
Untargeted Backdoor Watermark: Towards Harmless and Stealthy Dataset Copyright ProtectionNeural Information Processing Systems (NeurIPS), 2022
Yiming Li
Yang Bai
Yong Jiang
Yong-Liang Yang
Shutao Xia
Bo Li
AAML
412
137
0
27 Sep 2022
MOVE: Effective and Harmless Ownership Verification via Embedded External Features
MOVE: Effective and Harmless Ownership Verification via Embedded External FeaturesIEEE Transactions on Pattern Analysis and Machine Intelligence (TPAMI), 2022
Yiming Li
Linghui Zhu
Yang Liu
Yang Bai
Yong Jiang
Shutao Xia
Xiaochun Cao
Kui Ren
AAML
285
23
0
04 Aug 2022
Game of Trojans: A Submodular Byzantine Approach
Game of Trojans: A Submodular Byzantine Approach
D. Sahabandu
Arezoo Rajabi
Luyao Niu
Yangqiu Song
Bhaskar Ramasubramanian
Radha Poovendran
AAML
152
0
0
13 Jul 2022
Towards a Defense Against Federated Backdoor Attacks Under Continuous
  Training
Towards a Defense Against Federated Backdoor Attacks Under Continuous Training
Shuai Wang
J. Hayase
Giulia Fanti
Sewoong Oh
FedML
281
7
0
24 May 2022
Wild Patterns Reloaded: A Survey of Machine Learning Security against
  Training Data Poisoning
Wild Patterns Reloaded: A Survey of Machine Learning Security against Training Data PoisoningACM Computing Surveys (ACM CSUR), 2022
Antonio Emanuele Cinà
Kathrin Grosse
Ambra Demontis
Sebastiano Vascon
Werner Zellinger
Bernhard A. Moser
Alina Oprea
Battista Biggio
Marcello Pelillo
Fabio Roli
AAML
396
166
0
04 May 2022
Backdooring Explainable Machine Learning
Backdooring Explainable Machine Learning
Maximilian Noppel
Lukas Peter
Christian Wressnegger
AAML
202
5
0
20 Apr 2022
Trojan Horse Training for Breaking Defenses against Backdoor Attacks in
  Deep Learning
Trojan Horse Training for Breaking Defenses against Backdoor Attacks in Deep Learning
Arezoo Rajabi
Bhaskar Ramasubramanian
Radha Poovendran
AAML
208
5
0
25 Mar 2022
A Survey of Neural Trojan Attacks and Defenses in Deep Learning
A Survey of Neural Trojan Attacks and Defenses in Deep Learning
Jie Wang
Ghulam Mubashar Hassan
Naveed Akhtar
AAML
187
27
0
15 Feb 2022
Jigsaw Puzzle: Selective Backdoor Attack to Subvert Malware Classifiers
Jigsaw Puzzle: Selective Backdoor Attack to Subvert Malware ClassifiersIEEE Symposium on Security and Privacy (IEEE S&P), 2022
Limin Yang
Zhi Chen
Jacopo Cortellazzi
Feargus Pendlebury
Kevin Tu
Fabio Pierazzi
Lorenzo Cavallaro
Gang Wang
AAML
284
44
0
11 Feb 2022
Robust and Privacy-Preserving Collaborative Learning: A Comprehensive
  Survey
Robust and Privacy-Preserving Collaborative Learning: A Comprehensive Survey
Shangwei Guo
Xu Zhang
Feiyu Yang
Tianwei Zhang
Yan Gan
Tao Xiang
Yang Liu
FedML
231
13
0
19 Dec 2021
Towards Practical Deployment-Stage Backdoor Attack on Deep Neural
  Networks
Towards Practical Deployment-Stage Backdoor Attack on Deep Neural Networks
Xiangyu Qi
Tinghao Xie
Ruizhe Pan
Jifeng Zhu
Yong-Liang Yang
Kai Bu
AAML
248
73
0
25 Nov 2021
Poisoning Deep Reinforcement Learning Agents with In-Distribution
  Triggers
Poisoning Deep Reinforcement Learning Agents with In-Distribution Triggers
C. Ashcraft
Kiran Karra
151
31
0
14 Jun 2021
Stealthy Backdoors as Compression Artifacts
Stealthy Backdoors as Compression ArtifactsIEEE Transactions on Information Forensics and Security (IEEE TIFS), 2021
Yulong Tian
Fnu Suya
Fengyuan Xu
David Evans
228
29
0
30 Apr 2021
MISA: Online Defense of Trojaned Models using Misattributions
MISA: Online Defense of Trojaned Models using MisattributionsAsia-Pacific Computer Systems Architecture Conference (ACSA), 2021
Panagiota Kiourti
Wenchao Li
Anirban Roy
Karan Sikka
Susmit Jha
250
10
0
29 Mar 2021
Be Careful about Poisoned Word Embeddings: Exploring the Vulnerability
  of the Embedding Layers in NLP Models
Be Careful about Poisoned Word Embeddings: Exploring the Vulnerability of the Embedding Layers in NLP ModelsNorth American Chapter of the Association for Computational Linguistics (NAACL), 2021
Wenkai Yang
Lei Li
Zhiyuan Zhang
Xuancheng Ren
Xu Sun
Bin He
SILM
184
169
0
29 Mar 2021
Black-box Detection of Backdoor Attacks with Limited Information and
  Data
Black-box Detection of Backdoor Attacks with Limited Information and DataIEEE International Conference on Computer Vision (ICCV), 2021
Yinpeng Dong
Xiao Yang
Zhijie Deng
Tianyu Pang
Zihao Xiao
Hang Su
Jun Zhu
AAML
157
124
0
24 Mar 2021
EX-RAY: Distinguishing Injected Backdoor from Natural Features in Neural
  Networks by Examining Differential Feature Symmetry
EX-RAY: Distinguishing Injected Backdoor from Natural Features in Neural Networks by Examining Differential Feature Symmetry
Yingqi Liu
Guangyu Shen
Guanhong Tao
Zhenting Wang
Shiqing Ma
Xinming Zhang
AAML
224
8
0
16 Mar 2021
TrojanZoo: Towards Unified, Holistic, and Practical Evaluation of Neural
  Backdoors
TrojanZoo: Towards Unified, Holistic, and Practical Evaluation of Neural BackdoorsEuropean Symposium on Security and Privacy (EuroS&P), 2020
Ren Pang
Zheng Zhang
Xiangshan Gao
Zhaohan Xi
S. Ji
Peng Cheng
Xiapu Luo
Ting Wang
AAML
398
45
0
16 Dec 2020
Invisible Backdoor Attack with Sample-Specific Triggers
Invisible Backdoor Attack with Sample-Specific Triggers
Yuezun Li
Yiming Li
Baoyuan Wu
Longkang Li
Ran He
Siwei Lyu
AAMLDiffM
565
586
0
07 Dec 2020
Cassandra: Detecting Trojaned Networks from Adversarial Perturbations
Cassandra: Detecting Trojaned Networks from Adversarial PerturbationsIEEE Access (IEEE Access), 2020
Xiaoyu Zhang
Lin Wang
Rohit Gupta
Nazanin Rahnavard
M. Shah
AAML
196
27
0
28 Jul 2020
12
Next