ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1710.10547
  4. Cited By
Interpretation of Neural Networks is Fragile
v1v2 (latest)

Interpretation of Neural Networks is Fragile

AAAI Conference on Artificial Intelligence (AAAI), 2017
29 October 2017
Amirata Ghorbani
Abubakar Abid
James Zou
    FAttAAML
ArXiv (abs)PDFHTML

Papers citing "Interpretation of Neural Networks is Fragile"

50 / 489 papers shown
A Unified Concept-Based System for Local, Global, and Misclassification
  Explanations
A Unified Concept-Based System for Local, Global, and Misclassification ExplanationsNeurocomputing (Neurocomputing), 2023
Fatemeh Aghaeipoor
D. Asgarian
Mohammad Sabokrou
FAtt
192
0
0
06 Jun 2023
Adversarial attacks and defenses in explainable artificial intelligence: A survey
Adversarial attacks and defenses in explainable artificial intelligence: A surveyInformation Fusion (Inf. Fusion), 2023
Hubert Baniecki
P. Biecek
AAML
511
114
0
06 Jun 2023
Efficient Shapley Values Estimation by Amortization for Text
  Classification
Efficient Shapley Values Estimation by Amortization for Text ClassificationAnnual Meeting of the Association for Computational Linguistics (ACL), 2023
Chenghao Yang
Fan Yin
He He
Kai-Wei Chang
Xiaofei Ma
Bing Xiang
FAttVLM
243
7
0
31 May 2023
Are Your Explanations Reliable? Investigating the Stability of LIME in
  Explaining Text Classifiers by Marrying XAI and Adversarial Attack
Are Your Explanations Reliable? Investigating the Stability of LIME in Explaining Text Classifiers by Marrying XAI and Adversarial AttackConference on Empirical Methods in Natural Language Processing (EMNLP), 2023
Christopher Burger
Lingwei Chen
Thai Le
FAttAAML
232
18
0
21 May 2023
COCKATIEL: COntinuous Concept ranKed ATtribution with Interpretable
  ELements for explaining neural net classifiers on NLP tasks
COCKATIEL: COntinuous Concept ranKed ATtribution with Interpretable ELements for explaining neural net classifiers on NLP tasksAnnual Meeting of the Association for Computational Linguistics (ACL), 2023
Fanny Jourdan
Agustin Picard
Thomas Fel
Laurent Risser
Jean-Michel Loubes
Nicholas M. Asher
215
13
0
11 May 2023
Categorical Foundations of Explainable AI: A Unifying Theory
Categorical Foundations of Explainable AI: A Unifying Theory
Pietro Barbiero
S. Fioravanti
Francesco Giannini
Alberto Tonda
Pietro Lio
Elena Di Lavore
XAI
219
3
0
27 Apr 2023
Interpretable Neural-Symbolic Concept Reasoning
Interpretable Neural-Symbolic Concept ReasoningInternational Workshop on Neural-Symbolic Learning and Reasoning (NeSy), 2023
Pietro Barbiero
Gabriele Ciravegna
Francesco Giannini
M. Zarlenga
Lucie Charlotte Magister
Alberto Tonda
Pietro Lio
F. Precioso
M. Jamnik
G. Marra
NAILRM
378
55
0
27 Apr 2023
N$\text{A}^\text{2}$Q: Neural Attention Additive Model for Interpretable
  Multi-Agent Q-Learning
NA2\text{A}^\text{2}A2Q: Neural Attention Additive Model for Interpretable Multi-Agent Q-LearningInternational Conference on Machine Learning (ICML), 2023
Zichuan Liu
Yuanyang Zhu
Chunlin Chen
276
20
0
26 Apr 2023
An Efficient Ensemble Explainable AI (XAI) Approach for Morphed Face
  Detection
An Efficient Ensemble Explainable AI (XAI) Approach for Morphed Face DetectionPattern Recognition Letters (PR), 2023
Rudresh Dwivedi
Ritesh Kumar
Deepak Chopra
Pranay Kothari
Manjot Singh
CVBMAAML
158
12
0
23 Apr 2023
Explainability in AI Policies: A Critical Review of Communications,
  Reports, Regulations, and Standards in the EU, US, and UK
Explainability in AI Policies: A Critical Review of Communications, Reports, Regulations, and Standards in the EU, US, and UKConference on Fairness, Accountability and Transparency (FAccT), 2023
L. Nannini
Agathe Balayn
A. Smith
258
52
0
20 Apr 2023
Robustness of Visual Explanations to Common Data Augmentation
Robustness of Visual Explanations to Common Data Augmentation
Lenka Tětková
Lars Kai Hansen
AAML
213
6
0
18 Apr 2023
Evaluating the Robustness of Interpretability Methods through
  Explanation Invariance and Equivariance
Evaluating the Robustness of Interpretability Methods through Explanation Invariance and EquivarianceNeural Information Processing Systems (NeurIPS), 2023
Jonathan Crabbé
M. Schaar
AAML
251
13
0
13 Apr 2023
Optimizing Data Shapley Interaction Calculation from O(2^n) to O(t n^2)
  for KNN models
Optimizing Data Shapley Interaction Calculation from O(2^n) to O(t n^2) for KNN models
Mohamed Karim Belaid
Dorra El Mekki
Maximilian Rabus
Eyke Hüllermeier
170
3
0
02 Apr 2023
Foundation Models and Fair Use
Foundation Models and Fair UseJournal of machine learning research (JMLR), 2023
Peter Henderson
Xuechen Li
Dan Jurafsky
Tatsunori Hashimoto
Christopher De Sa
Abigail Z. Jacobs
176
156
0
28 Mar 2023
IDGI: A Framework to Eliminate Explanation Noise from Integrated
  Gradients
IDGI: A Framework to Eliminate Explanation Noise from Integrated GradientsComputer Vision and Pattern Recognition (CVPR), 2023
Ruo Yang
Binghui Wang
M. Bilgic
276
26
0
24 Mar 2023
Revisiting the Fragility of Influence Functions
Revisiting the Fragility of Influence FunctionsNeural Networks (Neural Netw.), 2023
Jacob R. Epifano
Ravichandran Ramachandran
A. Masino
Ghulam Rasool
TDI
275
22
0
22 Mar 2023
The Representational Status of Deep Learning Models
The Representational Status of Deep Learning Models
Eamon Duede
288
3
0
21 Mar 2023
It Is All About Data: A Survey on the Effects of Data on Adversarial
  Robustness
It Is All About Data: A Survey on the Effects of Data on Adversarial RobustnessACM Computing Surveys (ACM Comput. Surv.), 2023
Peiyu Xiong
Michael W. Tegegn
Jaskeerat Singh Sarin
Shubhraneel Pal
Julia Rubin
SILMAAML
370
16
0
17 Mar 2023
A Practical Upper Bound for the Worst-Case Attribution Deviations
A Practical Upper Bound for the Worst-Case Attribution DeviationsComputer Vision and Pattern Recognition (CVPR), 2023
Fan Wang
A. Kong
AAML
271
5
0
01 Mar 2023
SUNY: A Visual Interpretation Framework for Convolutional Neural
  Networks from a Necessary and Sufficient Perspective
SUNY: A Visual Interpretation Framework for Convolutional Neural Networks from a Necessary and Sufficient Perspective
Xiwei Xuan
Ziquan Deng
Hsuan-Tien Lin
Z. Kong
Kwan-Liu Ma
AAMLFAtt
327
5
0
01 Mar 2023
Don't be fooled: label leakage in explanation methods and the importance
  of their quantitative evaluation
Don't be fooled: label leakage in explanation methods and the importance of their quantitative evaluationInternational Conference on Artificial Intelligence and Statistics (AISTATS), 2023
Hao Zhang
A. Saporta
Rajesh Ranganath
FAtt
193
12
0
24 Feb 2023
SplineCam: Exact Visualization and Characterization of Deep Network
  Geometry and Decision Boundaries
SplineCam: Exact Visualization and Characterization of Deep Network Geometry and Decision BoundariesComputer Vision and Pattern Recognition (CVPR), 2023
Ahmed Imtiaz Humayun
Randall Balestriero
Guha Balakrishnan
Richard Baraniuk
321
27
0
24 Feb 2023
The Generalizability of Explanations
The Generalizability of ExplanationsIEEE International Joint Conference on Neural Network (IJCNN), 2023
Hanxiao Tan
FAtt
137
1
0
23 Feb 2023
sMRI-PatchNet: A novel explainable patch-based deep learning network for
  Alzheimer's disease diagnosis and discriminative atrophy localisation with
  Structural MRI
sMRI-PatchNet: A novel explainable patch-based deep learning network for Alzheimer's disease diagnosis and discriminative atrophy localisation with Structural MRIIEEE Access (IEEE Access), 2023
Xin Zhang
Liangxiu Han
Lianghao Han
Haoming Chen
Darren Dancey
Daoqiang Zhang
MedIm
263
17
0
17 Feb 2023
GCI: A (G)raph (C)oncept (I)nterpretation Framework
GCI: A (G)raph (C)oncept (I)nterpretation Framework
Dmitry Kazhdan
B. Dimanov
Lucie Charlotte Magister
Pietro Barbiero
M. Jamnik
Pietro Lio
162
5
0
09 Feb 2023
Diagnosing and Rectifying Vision Models using Language
Diagnosing and Rectifying Vision Models using LanguageInternational Conference on Learning Representations (ICLR), 2023
Yuhui Zhang
Jeff Z. HaoChen
Shih-Cheng Huang
Kuan-Chieh Wang
James Zou
Serena Yeung
206
61
0
08 Feb 2023
Certified Interpretability Robustness for Class Activation Mapping
Certified Interpretability Robustness for Class Activation Mapping
Alex Gu
Tsui-Wei Weng
Pin-Yu Chen
Sijia Liu
Lucani E. Daniel
AAML
141
2
0
26 Jan 2023
Explainable AI does not provide the explanations end-users are asking
  for
Explainable AI does not provide the explanations end-users are asking for
Savio Rozario
G. Cevora
XAI
186
2
0
25 Jan 2023
MoreauGrad: Sparse and Robust Interpretation of Neural Networks via
  Moreau Envelope
MoreauGrad: Sparse and Robust Interpretation of Neural Networks via Moreau EnvelopeIEEE International Conference on Computer Vision (ICCV), 2023
Jingwei Zhang
Farzan Farnia
UQCV
142
4
0
08 Jan 2023
Valid P-Value for Deep Learning-Driven Salient Region
Valid P-Value for Deep Learning-Driven Salient RegionInternational Conference on Learning Representations (ICLR), 2023
Daiki Miwa
Vo Nguyen Le Duy
I. Takeuchi
FAttAAML
188
18
0
06 Jan 2023
PEAK: Explainable Privacy Assistant through Automated Knowledge
  Extraction
PEAK: Explainable Privacy Assistant through Automated Knowledge Extraction
Gonul Ayci
Arzucan Özgür
Murat cSensoy
P. Yolum
244
4
0
05 Jan 2023
Disentangled Explanations of Neural Network Predictions by Finding
  Relevant Subspaces
Disentangled Explanations of Neural Network Predictions by Finding Relevant SubspacesIEEE Transactions on Pattern Analysis and Machine Intelligence (TPAMI), 2022
Pattarawat Chormai
J. Herrmann
Klaus-Robert Muller
G. Montavon
FAtt
420
27
0
30 Dec 2022
Provable Robust Saliency-based Explanations
Provable Robust Saliency-based Explanations
Chao Chen
Chenghua Guo
Guixiang Ma
Ming Zeng
Xi Zhang
Sihong Xie
AAMLFAtt
413
1
0
28 Dec 2022
On the Equivalence of the Weighted Tsetlin Machine and the Perceptron
On the Equivalence of the Weighted Tsetlin Machine and the Perceptron
Jivitesh Sharma
Ole-Christoffer Granmo
Lei Jiao
163
1
0
27 Dec 2022
The Quantum Path Kernel: a Generalized Quantum Neural Tangent Kernel for
  Deep Quantum Machine Learning
The Quantum Path Kernel: a Generalized Quantum Neural Tangent Kernel for Deep Quantum Machine LearningIEEE Transactions on Quantum Engineering (IEEE Trans. Quantum Eng.), 2022
Massimiliano Incudini
Michele Grossi
Antonio Mandarino
S. Vallecorsa
Alessandra Di Pierro
David Windridge
266
14
0
22 Dec 2022
AI Security for Geoscience and Remote Sensing: Challenges and Future
  Trends
AI Security for Geoscience and Remote Sensing: Challenges and Future TrendsIEEE Geoscience and Remote Sensing Magazine (GRSM), 2022
Yonghao Xu
Tao Bai
Weikang Yu
Shizhen Chang
P. M. Atkinson
Pedram Ghamisi
AAML
308
61
0
19 Dec 2022
Estimating the Adversarial Robustness of Attributions in Text with
  Transformers
Estimating the Adversarial Robustness of Attributions in Text with Transformers
Adam Ivankay
Mattia Rigotti
Ivan Girardi
Chiara Marchiori
P. Frossard
149
1
0
18 Dec 2022
Robust Explanation Constraints for Neural Networks
Robust Explanation Constraints for Neural NetworksInternational Conference on Learning Representations (ICLR), 2022
Matthew Wicker
Juyeon Heo
Luca Costabello
Adrian Weller
FAtt
235
23
0
16 Dec 2022
Interpretable ML for Imbalanced Data
Interpretable ML for Imbalanced Data
Damien Dablain
C. Bellinger
Bartosz Krawczyk
D. Aha
Nitesh Chawla
202
2
0
15 Dec 2022
Identifying the Source of Vulnerability in Explanation Discrepancy: A
  Case Study in Neural Text Classification
Identifying the Source of Vulnerability in Explanation Discrepancy: A Case Study in Neural Text ClassificationBlackboxNLP Workshop on Analyzing and Interpreting Neural Networks for NLP (BlackboxNLP), 2022
Ruixuan Tang
Hanjie Chen
Yangfeng Ji
AAMLFAtt
186
3
0
10 Dec 2022
Spurious Features Everywhere -- Large-Scale Detection of Harmful
  Spurious Features in ImageNet
Spurious Features Everywhere -- Large-Scale Detection of Harmful Spurious Features in ImageNetIEEE International Conference on Computer Vision (ICCV), 2022
Yannic Neuhaus
Maximilian Augustin
Valentyn Boreiko
Matthias Hein
AAML
299
39
0
09 Dec 2022
Post hoc Explanations may be Ineffective for Detecting Unknown Spurious
  Correlation
Post hoc Explanations may be Ineffective for Detecting Unknown Spurious CorrelationInternational Conference on Learning Representations (ICLR), 2022
Julius Adebayo
M. Muelly
H. Abelson
Been Kim
233
93
0
09 Dec 2022
XRand: Differentially Private Defense against Explanation-Guided Attacks
XRand: Differentially Private Defense against Explanation-Guided AttacksAAAI Conference on Artificial Intelligence (AAAI), 2022
Truc D. T. Nguyen
Phung Lai
Nhathai Phan
My T. Thai
AAMLSILM
298
20
0
08 Dec 2022
Interpretation of Neural Networks is Susceptible to Universal
  Adversarial Perturbations
Interpretation of Neural Networks is Susceptible to Universal Adversarial PerturbationsIEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP), 2022
Haniyeh Ehsani Oskouie
Farzan Farnia
FAttAAML
267
6
0
30 Nov 2022
Understanding and Enhancing Robustness of Concept-based Models
Understanding and Enhancing Robustness of Concept-based ModelsAAAI Conference on Artificial Intelligence (AAAI), 2022
Sanchit Sinha
Mengdi Huai
Jianhui Sun
Aidong Zhang
AAML
144
27
0
29 Nov 2022
Towards More Robust Interpretation via Local Gradient Alignment
Towards More Robust Interpretation via Local Gradient AlignmentAAAI Conference on Artificial Intelligence (AAAI), 2022
Sunghwan Joo
Seokhyeon Jeong
Juyeon Heo
Adrian Weller
Taesup Moon
FAtt
245
7
0
29 Nov 2022
Foiling Explanations in Deep Neural Networks
Foiling Explanations in Deep Neural Networks
Snir Vitrack Tamam
Raz Lapid
Moshe Sipper
AAML
207
21
0
27 Nov 2022
SEAT: Stable and Explainable Attention
SEAT: Stable and Explainable AttentionAAAI Conference on Artificial Intelligence (AAAI), 2022
Lijie Hu
Yixin Liu
Ninghao Liu
Mengdi Huai
Lichao Sun
Haiyan Zhao
OOD
152
25
0
23 Nov 2022
Concept-based Explanations using Non-negative Concept Activation Vectors
  and Decision Tree for CNN Models
Concept-based Explanations using Non-negative Concept Activation Vectors and Decision Tree for CNN Models
Gayda Mutahar
Tim Miller
FAtt
153
8
0
19 Nov 2022
Data-Adaptive Discriminative Feature Localization with Statistically
  Guaranteed Interpretation
Data-Adaptive Discriminative Feature Localization with Statistically Guaranteed InterpretationAnnals of Applied Statistics (AOAS), 2022
Ben Dai
Xiaotong Shen
Lingzhi Chen
Chunlin Li
Wei Pan
FAtt
145
2
0
18 Nov 2022
Previous
12345...8910
Next