ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1710.10547
  4. Cited By
Interpretation of Neural Networks is Fragile
v1v2 (latest)

Interpretation of Neural Networks is Fragile

AAAI Conference on Artificial Intelligence (AAAI), 2017
29 October 2017
Amirata Ghorbani
Abubakar Abid
James Zou
    FAttAAML
ArXiv (abs)PDFHTML

Papers citing "Interpretation of Neural Networks is Fragile"

50 / 489 papers shown
DeepAID: Interpreting and Improving Deep Learning-based Anomaly
  Detection in Security Applications
DeepAID: Interpreting and Improving Deep Learning-based Anomaly Detection in Security ApplicationsConference on Computer and Communications Security (CCS), 2021
Dongqi Han
Zhiliang Wang
Wenqi Chen
Ying Zhong
Su Wang
Han Zhang
Jiahai Yang
Xingang Shi
Xia Yin
AAML
166
105
0
23 Sep 2021
Ranking Feature-Block Importance in Artificial Multiblock Neural
  Networks
Ranking Feature-Block Importance in Artificial Multiblock Neural Networks
Anna Jenul
Stefan Schrunner
B. Huynh
Runar Helin
C. Futsaether
K. H. Liland
O. Tomic
FAtt
133
1
0
21 Sep 2021
FUTURE-AI: Guiding Principles and Consensus Recommendations for
  Trustworthy Artificial Intelligence in Medical Imaging
FUTURE-AI: Guiding Principles and Consensus Recommendations for Trustworthy Artificial Intelligence in Medical Imaging
Karim Lekadira
Richard Osuala
C. Gallin
Noussair Lazrak
Kaisar Kushibar
...
Nickolas Papanikolaou
Zohaib Salahuddin
Henry C. Woodruff
Philippe Lambin
L. Martí-Bonmatí
AI4TS
358
79
0
20 Sep 2021
Self-learn to Explain Siamese Networks Robustly
Self-learn to Explain Siamese Networks Robustly
Chao Chen
Yifan Shen
Guixiang Ma
Xiangnan Kong
S. Rangarajan
Xi Zhang
Sihong Xie
163
7
0
15 Sep 2021
Rationales for Sequential Predictions
Rationales for Sequential Predictions
Keyon Vafa
Yuntian Deng
David M. Blei
Alexander M. Rush
218
36
0
14 Sep 2021
Logic Traps in Evaluating Attribution Scores
Logic Traps in Evaluating Attribution Scores
Yiming Ju
Yuanzhe Zhang
Zhao Yang
Zhongtao Jiang
Kang Liu
Jun Zhao
XAIFAtt
276
23
0
12 Sep 2021
EG-Booster: Explanation-Guided Booster of ML Evasion Attacks
EG-Booster: Explanation-Guided Booster of ML Evasion AttacksConference on Data and Application Security and Privacy (CODASPY), 2021
Abderrahmen Amich
Birhanu Eshete
AAML
146
10
0
31 Aug 2021
Enjoy the Salience: Towards Better Transformer-based Faithful
  Explanations with Word Salience
Enjoy the Salience: Towards Better Transformer-based Faithful Explanations with Word SalienceConference on Empirical Methods in Natural Language Processing (EMNLP), 2021
G. Chrysostomou
Nikolaos Aletras
197
22
0
31 Aug 2021
Finding Representative Interpretations on Convolutional Neural Networks
Finding Representative Interpretations on Convolutional Neural NetworksIEEE International Conference on Computer Vision (ICCV), 2021
P. C. Lam
Lingyang Chu
Maxim Torgonskiy
Jian Pei
Yong Zhang
Lanjun Wang
FAttSSLHAI
175
7
0
13 Aug 2021
Jujutsu: A Two-stage Defense against Adversarial Patch Attacks on Deep
  Neural Networks
Jujutsu: A Two-stage Defense against Adversarial Patch Attacks on Deep Neural NetworksACM Asia Conference on Computer and Communications Security (AsiaCCS), 2021
Zitao Chen
Pritam Dash
Karthik Pattabiraman
AAML
339
27
0
11 Aug 2021
Perturbing Inputs for Fragile Interpretations in Deep Natural Language
  Processing
Perturbing Inputs for Fragile Interpretations in Deep Natural Language ProcessingBlackboxNLP Workshop on Analyzing and Interpreting Neural Networks for NLP (BlackBoxNLP), 2021
Sanchit Sinha
Hanjie Chen
Arshdeep Sekhon
Yangfeng Ji
Yanjun Qi
AAMLFAtt
249
48
0
11 Aug 2021
Harnessing value from data science in business: ensuring explainability
  and fairness of solutions
Harnessing value from data science in business: ensuring explainability and fairness of solutions
Krzysztof Chomiak
Michal Miktus
97
0
0
10 Aug 2021
Explainable AI and susceptibility to adversarial attacks: a case study
  in classification of breast ultrasound images
Explainable AI and susceptibility to adversarial attacks: a case study in classification of breast ultrasound imagesIUS (IUS), 2021
Hamza Rasaee
H. Rivaz
AAML
99
21
0
09 Aug 2021
Jointly Attacking Graph Neural Network and its Explanations
Jointly Attacking Graph Neural Network and its ExplanationsIEEE International Conference on Data Engineering (ICDE), 2021
Wenqi Fan
Wei Jin
Xiaorui Liu
Han Xu
Xianfeng Tang
Suhang Wang
Qing Li
Shucheng Zhou
Jianping Wang
Charu C. Aggarwal
AAML
246
33
0
07 Aug 2021
Resisting Out-of-Distribution Data Problem in Perturbation of XAI
Resisting Out-of-Distribution Data Problem in Perturbation of XAI
Luyu Qiu
Yi Yang
Caleb Chen Cao
Jing Liu
Yueyuan Zheng
H. Ngai
J. H. Hsiao
Lei Chen
235
19
0
27 Jul 2021
Robust Explainability: A Tutorial on Gradient-Based Attribution Methods
  for Deep Neural Networks
Robust Explainability: A Tutorial on Gradient-Based Attribution Methods for Deep Neural NetworksIEEE Signal Processing Magazine (IEEE SPM), 2021
Ian E. Nielsen
Dimah Dera
Ghulam Rasool
N. Bouaynaya
R. Ramachandran
FAtt
359
101
0
23 Jul 2021
Trustworthy AI: A Computational Perspective
Trustworthy AI: A Computational Perspective
Haochen Liu
Yiqi Wang
Wenqi Fan
Xiaorui Liu
Yaxin Li
Shaili Jain
Yunhao Liu
Anil K. Jain
Shucheng Zhou
FaML
399
256
0
12 Jul 2021
Robust Counterfactual Explanations on Graph Neural Networks
Robust Counterfactual Explanations on Graph Neural NetworksNeural Information Processing Systems (NeurIPS), 2021
Mohit Bajaj
Lingyang Chu
Zihui Xue
Jian Pei
Lanjun Wang
P. C. Lam
Yong Zhang
OOD
406
114
0
08 Jul 2021
When and How to Fool Explainable Models (and Humans) with Adversarial
  Examples
When and How to Fool Explainable Models (and Humans) with Adversarial Examples
Jon Vadillo
Roberto Santana
Jose A. Lozano
SILMAAML
261
21
0
05 Jul 2021
Certifiably Robust Interpretation via Renyi Differential Privacy
Certifiably Robust Interpretation via Renyi Differential Privacy
Ao Liu
Xiaoyu Chen
Sijia Liu
Lirong Xia
Chuang Gan
AAML
143
14
0
04 Jul 2021
Explanation-Guided Diagnosis of Machine Learning Evasion Attacks
Explanation-Guided Diagnosis of Machine Learning Evasion AttacksSecurity and Privacy in Communication Networks (SecureComm), 2021
Abderrahmen Amich
Birhanu Eshete
AAML
115
14
0
30 Jun 2021
On Locality of Local Explanation Models
On Locality of Local Explanation Models
Sahra Ghalebikesabi
Lucile Ter-Minassian
Karla Diaz-Ordaz
Chris Holmes
FedMLFAtt
155
44
0
24 Jun 2021
Guided Integrated Gradients: An Adaptive Path Method for Removing Noise
Guided Integrated Gradients: An Adaptive Path Method for Removing NoiseComputer Vision and Pattern Recognition (CVPR), 2021
A. Kapishnikov
Subhashini Venugopalan
Besim Avci
Benjamin D. Wedin
Michael Terry
Tolga Bolukbasi
263
123
0
17 Jun 2021
Best of both worlds: local and global explanations with
  human-understandable concepts
Best of both worlds: local and global explanations with human-understandable concepts
Jessica Schrouff
Sebastien Baur
Shaobo Hou
Diana Mincu
Eric Loreaux
Ralph Blanes
James Wexler
Alan Karthikesalingam
Been Kim
FAtt
226
31
0
16 Jun 2021
S-LIME: Stabilized-LIME for Model Explanation
S-LIME: Stabilized-LIME for Model ExplanationKnowledge Discovery and Data Mining (KDD), 2021
Zhengze Zhou
Giles Hooker
Fei Wang
FAtt
253
113
0
15 Jun 2021
On the Lack of Robust Interpretability of Neural Text Classifiers
On the Lack of Robust Interpretability of Neural Text ClassifiersFindings (Findings), 2021
Muhammad Bilal Zafar
Michele Donini
Dylan Slack
Cédric Archambeau
Sanjiv Ranjan Das
K. Kenthapadi
AAML
119
23
0
08 Jun 2021
3DB: A Framework for Debugging Computer Vision Models
3DB: A Framework for Debugging Computer Vision ModelsNeural Information Processing Systems (NeurIPS), 2021
Guillaume Leclerc
Hadi Salman
Andrew Ilyas
Sai H. Vemprala
Logan Engstrom
...
Pengchuan Zhang
Shibani Santurkar
Greg Yang
Ashish Kapoor
Aleksander Madry
240
44
0
07 Jun 2021
Evaluating Local Explanations using White-box Models
Evaluating Local Explanations using White-box Models
Amir Hossein Akhavan Rahnama
Judith Butepage
Pierre Geurts
Henrik Bostrom
FAtt
202
0
0
04 Jun 2021
DISSECT: Disentangled Simultaneous Explanations via Concept Traversals
DISSECT: Disentangled Simultaneous Explanations via Concept TraversalsInternational Conference on Learning Representations (ICLR), 2021
Asma Ghandeharioun
Been Kim
Chun-Liang Li
Brendan Jou
B. Eoff
Rosalind W. Picard
AAML
328
57
0
31 May 2021
The effectiveness of feature attribution methods and its correlation
  with automatic evaluation scores
The effectiveness of feature attribution methods and its correlation with automatic evaluation scoresNeural Information Processing Systems (NeurIPS), 2021
Giang Nguyen
Daeyoung Kim
Anh Totti Nguyen
FAtt
482
105
0
31 May 2021
Drop Clause: Enhancing Performance, Interpretability and Robustness of
  the Tsetlin Machine
Drop Clause: Enhancing Performance, Interpretability and Robustness of the Tsetlin Machine
Jivitesh Sharma
Rohan Kumar Yadav
Ole-Christoffer Granmo
Lei Jiao
VLM
206
13
0
30 May 2021
EDDA: Explanation-driven Data Augmentation to Improve Explanation
  Faithfulness
EDDA: Explanation-driven Data Augmentation to Improve Explanation Faithfulness
Ruiwen Li
Zhibo Zhang
Jiani Li
C. Trabelsi
Scott Sanner
Jongseong Jang
Yeonjeong Jeong
Dongsub Shim
AAML
212
1
0
29 May 2021
Fooling Partial Dependence via Data Poisoning
Fooling Partial Dependence via Data Poisoning
Hubert Baniecki
Wojciech Kretowicz
P. Biecek
AAML
296
28
0
26 May 2021
Information-theoretic Evolution of Model Agnostic Global Explanations
Information-theoretic Evolution of Model Agnostic Global Explanations
Sukriti Verma
Nikaash Puri
Piyush B. Gupta
Balaji Krishnamurthy
FAtt
173
0
0
14 May 2021
XAI Handbook: Towards a Unified Framework for Explainable AI
XAI Handbook: Towards a Unified Framework for Explainable AI
Sebastián M. Palacio
Adriano Lucieri
Mohsin Munir
Jörn Hees
Sheraz Ahmed
Andreas Dengel
133
40
0
14 May 2021
Leveraging Sparse Linear Layers for Debuggable Deep Networks
Leveraging Sparse Linear Layers for Debuggable Deep NetworksInternational Conference on Machine Learning (ICML), 2021
Eric Wong
Shibani Santurkar
Aleksander Madry
FAtt
208
96
0
11 May 2021
Interpretable Semantic Photo Geolocation
Interpretable Semantic Photo GeolocationIEEE Workshop/Winter Conference on Applications of Computer Vision (WACV), 2021
Jonas Theiner
Eric Müller-Budack
Ralph Ewerth
181
41
0
30 Apr 2021
Towards Adversarial Patch Analysis and Certified Defense against Crowd
  Counting
Towards Adversarial Patch Analysis and Certified Defense against Crowd CountingACM Multimedia (ACM MM), 2021
Qiming Wu
Zhikang Zou
Pan Zhou
Xiaoqing Ye
Binghui Wang
Ang Li
AAML
246
7
0
22 Apr 2021
On the Sensitivity and Stability of Model Interpretations in NLP
On the Sensitivity and Stability of Model Interpretations in NLPAnnual Meeting of the Association for Computational Linguistics (ACL), 2021
Fan Yin
Zhouxing Shi
Cho-Jui Hsieh
Kai-Wei Chang
FAtt
248
34
0
18 Apr 2021
Evaluating Saliency Methods for Neural Language Models
Evaluating Saliency Methods for Neural Language ModelsNorth American Chapter of the Association for Computational Linguistics (NAACL), 2021
Shuoyang Ding
Philipp Koehn
FAttXAI
122
60
0
12 Apr 2021
A-FMI: Learning Attributions from Deep Networks via Feature Map
  Importance
A-FMI: Learning Attributions from Deep Networks via Feature Map Importance
An Zhang
Xiang Wang
Chengfang Fang
Jie Shi
Tat-Seng Chua
Zehua Chen
FAtt
87
3
0
12 Apr 2021
Sparse Oblique Decision Trees: A Tool to Understand and Manipulate
  Neural Net Features
Sparse Oblique Decision Trees: A Tool to Understand and Manipulate Neural Net FeaturesData mining and knowledge discovery (DMKD), 2021
Suryabhan Singh Hada
Miguel Á. Carreira-Perpiñán
Arman Zharmagambetov
199
19
0
07 Apr 2021
Neural Response Interpretation through the Lens of Critical Pathways
Neural Response Interpretation through the Lens of Critical PathwaysComputer Vision and Pattern Recognition (CVPR), 2021
Ashkan Khakzar
Soroosh Baselizadeh
Saurabh Khanduja
Christian Rupprecht
Seong Tae Kim
Nassir Navab
123
41
0
31 Mar 2021
Building Reliable Explanations of Unreliable Neural Networks: Locally
  Smoothing Perspective of Model Interpretation
Building Reliable Explanations of Unreliable Neural Networks: Locally Smoothing Perspective of Model InterpretationComputer Vision and Pattern Recognition (CVPR), 2021
Dohun Lim
Hyeonseok Lee
Sungchan Kim
FAttAAML
160
13
0
26 Mar 2021
ExAD: An Ensemble Approach for Explanation-based Adversarial Detection
ExAD: An Ensemble Approach for Explanation-based Adversarial Detection
R. Vardhan
Ninghao Liu
Phakpoom Chinprutthiwong
Weijie Fu
Zhen Hu
Helen Zhou
G. Gu
AAML
193
5
0
22 Mar 2021
CACTUS: Detecting and Resolving Conflicts in Objective Functions
CACTUS: Detecting and Resolving Conflicts in Objective Functions
Subhajit Das
Alex Endert
95
0
0
13 Mar 2021
Human-Understandable Decision Making for Visual Recognition
Human-Understandable Decision Making for Visual RecognitionPacific-Asia Conference on Knowledge Discovery and Data Mining (PAKDD), 2021
Xiaowei Zhou
Jie Yin
Ivor Tsang
Chen Wang
FAttHAI
122
1
0
05 Mar 2021
Detecting Spurious Correlations with Sanity Tests for Artificial
  Intelligence Guided Radiology Systems
Detecting Spurious Correlations with Sanity Tests for Artificial Intelligence Guided Radiology SystemsFrontiers in Digital Health (FDH), 2021
U. Mahmood
Robik Shrestha
D. Bates
L. Mannelli
G. Corrias
Y. Erdi
Christopher Kanan
183
20
0
04 Mar 2021
Do Input Gradients Highlight Discriminative Features?
Do Input Gradients Highlight Discriminative Features?Neural Information Processing Systems (NeurIPS), 2021
Harshay Shah
Prateek Jain
Praneeth Netrapalli
AAMLFAtt
362
66
0
25 Feb 2021
Resilience of Bayesian Layer-Wise Explanations under Adversarial Attacks
Resilience of Bayesian Layer-Wise Explanations under Adversarial AttacksIEEE International Joint Conference on Neural Network (IJCNN), 2021
Ginevra Carbone
G. Sanguinetti
Luca Bortolussi
FAttAAML
279
4
0
22 Feb 2021
Previous
123...106789
Next