ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1710.10547
  4. Cited By
Interpretation of Neural Networks is Fragile

Interpretation of Neural Networks is Fragile

29 October 2017
Amirata Ghorbani
Abubakar Abid
James Y. Zou
    FAtt
    AAML
ArXivPDFHTML

Papers citing "Interpretation of Neural Networks is Fragile"

50 / 467 papers shown
Title
Faithful Explanations of Black-box NLP Models Using LLM-generated
  Counterfactuals
Faithful Explanations of Black-box NLP Models Using LLM-generated Counterfactuals
Y. Gat
Nitay Calderon
Amir Feder
Alexander Chapanin
Amit Sharma
Roi Reichart
35
28
0
01 Oct 2023
Black-box Attacks on Image Activity Prediction and its Natural Language
  Explanations
Black-box Attacks on Image Activity Prediction and its Natural Language Explanations
Alina Elena Baia
Valentina Poggioni
Andrea Cavallaro
AAML
16
1
0
30 Sep 2023
DeepROCK: Error-controlled interaction detection in deep neural networks
DeepROCK: Error-controlled interaction detection in deep neural networks
Winston Chen
William Stafford Noble
Y. Lu
6
1
0
26 Sep 2023
Concept explainability for plant diseases classification
Concept explainability for plant diseases classification
Jihen Amara
B. König-Ries
Sheeba Samuel
FAtt
11
2
0
15 Sep 2023
Interpretability is in the Mind of the Beholder: A Causal Framework for
  Human-interpretable Representation Learning
Interpretability is in the Mind of the Beholder: A Causal Framework for Human-interpretable Representation Learning
Emanuele Marconato
Andrea Passerini
Stefano Teso
30
13
0
14 Sep 2023
Automatic Concept Embedding Model (ACEM): No train-time concepts, No
  issue!
Automatic Concept Embedding Model (ACEM): No train-time concepts, No issue!
Rishabh Jain
LRM
27
0
0
07 Sep 2023
Goodhart's Law Applies to NLP's Explanation Benchmarks
Goodhart's Law Applies to NLP's Explanation Benchmarks
Jennifer Hsia
Danish Pruthi
Aarti Singh
Zachary Chase Lipton
28
6
0
28 Aug 2023
On the Interpretability of Quantum Neural Networks
On the Interpretability of Quantum Neural Networks
Lirande Pira
C. Ferrie
FAtt
27
15
0
22 Aug 2023
FUTURE-AI: International consensus guideline for trustworthy and
  deployable artificial intelligence in healthcare
FUTURE-AI: International consensus guideline for trustworthy and deployable artificial intelligence in healthcare
Karim Lekadir
Aasa Feragen
Abdul Joseph Fofanah
Alejandro F Frangi
Alena Buyx
...
Yi Zeng
Yunusa G Mohammed
Yves Saint James Aquino
Zohaib Salahuddin
M. P. Starmans
AI4TS
32
38
0
11 Aug 2023
A New Perspective on Evaluation Methods for Explainable Artificial
  Intelligence (XAI)
A New Perspective on Evaluation Methods for Explainable Artificial Intelligence (XAI)
Timo Speith
Markus Langer
26
12
0
26 Jul 2023
Saliency strikes back: How filtering out high frequencies improves
  white-box explanations
Saliency strikes back: How filtering out high frequencies improves white-box explanations
Sabine Muzellec
Thomas Fel
Victor Boutin
Léo Andéol
R. V. Rullen
Thomas Serre
FAtt
22
0
0
18 Jul 2023
SHAMSUL: Systematic Holistic Analysis to investigate Medical
  Significance Utilizing Local interpretability methods in deep learning for
  chest radiography pathology prediction
SHAMSUL: Systematic Holistic Analysis to investigate Medical Significance Utilizing Local interpretability methods in deep learning for chest radiography pathology prediction
Mahbub Ul Alam
Jaakko Hollmén
Jón R. Baldvinsson
R. Rahmani
FAtt
31
1
0
16 Jul 2023
On the Connection between Game-Theoretic Feature Attributions and
  Counterfactual Explanations
On the Connection between Game-Theoretic Feature Attributions and Counterfactual Explanations
Emanuele Albini
Shubham Sharma
Saumitra Mishra
Danial Dervovic
Daniele Magazzeni
FAtt
46
2
0
13 Jul 2023
Single-Class Target-Specific Attack against Interpretable Deep Learning
  Systems
Single-Class Target-Specific Attack against Interpretable Deep Learning Systems
Eldor Abdukhamidov
Mohammed Abuhamad
George K. Thiruvathukal
Hyoungshick Kim
Tamer Abuhmed
AAML
25
2
0
12 Jul 2023
Stability Guarantees for Feature Attributions with Multiplicative
  Smoothing
Stability Guarantees for Feature Attributions with Multiplicative Smoothing
Anton Xue
Rajeev Alur
Eric Wong
36
5
0
12 Jul 2023
Hierarchical Semantic Tree Concept Whitening for Interpretable Image
  Classification
Hierarchical Semantic Tree Concept Whitening for Interpretable Image Classification
Haixing Dai
Lu Zhang
Lin Zhao
Zihao Wu
Zheng Liu
...
Yanjun Lyu
Changying Li
Ninghao Liu
Tianming Liu
Dajiang Zhu
43
6
0
10 Jul 2023
Robust Ranking Explanations
Robust Ranking Explanations
Chao Chen
Chenghua Guo
Guixiang Ma
Ming Zeng
Xi Zhang
Sihong Xie
FAtt
AAML
35
0
0
08 Jul 2023
A Vulnerability of Attribution Methods Using Pre-Softmax Scores
A Vulnerability of Attribution Methods Using Pre-Softmax Scores
Miguel A. Lerma
Mirtha Lucas
FAtt
19
0
0
06 Jul 2023
DARE: Towards Robust Text Explanations in Biomedical and Healthcare
  Applications
DARE: Towards Robust Text Explanations in Biomedical and Healthcare Applications
Adam Ivankay
Mattia Rigotti
P. Frossard
OOD
MedIm
21
1
0
05 Jul 2023
Fixing confirmation bias in feature attribution methods via semantic
  match
Fixing confirmation bias in feature attribution methods via semantic match
Giovanni Cina
Daniel Fernandez-Llaneza
Ludovico Deponte
Nishant Mishra
Tabea E. Rober
Sandro Pezzelle
Iacer Calixto
Rob Goedhart
cS. .Ilker Birbil
FAtt
22
0
0
03 Jul 2023
SHARCS: Shared Concept Space for Explainable Multimodal Learning
SHARCS: Shared Concept Space for Explainable Multimodal Learning
Gabriele Dominici
Pietro Barbiero
Lucie Charlotte Magister
Pietro Lio'
Nikola Simidjievski
23
5
0
01 Jul 2023
Verifying Safety of Neural Networks from Topological Perspectives
Verifying Safety of Neural Networks from Topological Perspectives
Zhen Liang
Dejin Ren
Bai Xue
J. Wang
Wenjing Yang
Wanwei Liu
AAML
30
0
0
27 Jun 2023
Requirements for Explainability and Acceptance of Artificial
  Intelligence in Collaborative Work
Requirements for Explainability and Acceptance of Artificial Intelligence in Collaborative Work
Sabine Theis
Sophie F. Jentzsch
Fotini Deligiannaki
C. Berro
A. Raulf
C. Bruder
20
8
0
27 Jun 2023
Evaluating the overall sensitivity of saliency-based explanation methods
Evaluating the overall sensitivity of saliency-based explanation methods
Harshinee Sriram
Cristina Conati
AAML
XAI
FAtt
16
0
0
21 Jun 2023
On the Robustness of Removal-Based Feature Attributions
On the Robustness of Removal-Based Feature Attributions
Christy Lin
Ian Covert
Su-In Lee
20
12
0
12 Jun 2023
A Holistic Approach to Unifying Automatic Concept Extraction and Concept
  Importance Estimation
A Holistic Approach to Unifying Automatic Concept Extraction and Concept Importance Estimation
Thomas Fel
Victor Boutin
Mazda Moayeri
Rémi Cadène
Louis Bethune
Léo Andéol
Mathieu Chalvidal
Thomas Serre
FAtt
16
49
0
11 Jun 2023
On Minimizing the Impact of Dataset Shifts on Actionable Explanations
On Minimizing the Impact of Dataset Shifts on Actionable Explanations
Anna P. Meyer
Dan Ley
Suraj Srinivas
Himabindu Lakkaraju
FAtt
34
6
0
11 Jun 2023
Adversarial attacks and defenses in explainable artificial intelligence:
  A survey
Adversarial attacks and defenses in explainable artificial intelligence: A survey
Hubert Baniecki
P. Biecek
AAML
42
63
0
06 Jun 2023
A Unified Concept-Based System for Local, Global, and Misclassification
  Explanations
A Unified Concept-Based System for Local, Global, and Misclassification Explanations
Fatemeh Aghaeipoor
D. Asgarian
Mohammad Sabokrou
FAtt
19
0
0
06 Jun 2023
Efficient Shapley Values Estimation by Amortization for Text
  Classification
Efficient Shapley Values Estimation by Amortization for Text Classification
Chenghao Yang
Fan Yin
He He
Kai-Wei Chang
Xiaofei Ma
Bing Xiang
FAtt
VLM
18
4
0
31 May 2023
Are Your Explanations Reliable? Investigating the Stability of LIME in
  Explaining Text Classifiers by Marrying XAI and Adversarial Attack
Are Your Explanations Reliable? Investigating the Stability of LIME in Explaining Text Classifiers by Marrying XAI and Adversarial Attack
Christopher Burger
Lingwei Chen
Thai Le
FAtt
AAML
22
9
0
21 May 2023
COCKATIEL: COntinuous Concept ranKed ATtribution with Interpretable
  ELements for explaining neural net classifiers on NLP tasks
COCKATIEL: COntinuous Concept ranKed ATtribution with Interpretable ELements for explaining neural net classifiers on NLP tasks
Fanny Jourdan
Agustin Picard
Thomas Fel
Laurent Risser
Jean-Michel Loubes
Nicholas M. Asher
15
6
0
11 May 2023
Categorical Foundations of Explainable AI: A Unifying Theory
Categorical Foundations of Explainable AI: A Unifying Theory
Pietro Barbiero
S. Fioravanti
Francesco Giannini
Alberto Tonda
Pietro Lio'
Elena Di Lavore
XAI
19
2
0
27 Apr 2023
Interpretable Neural-Symbolic Concept Reasoning
Interpretable Neural-Symbolic Concept Reasoning
Pietro Barbiero
Gabriele Ciravegna
Francesco Giannini
M. Zarlenga
Lucie Charlotte Magister
Alberto Tonda
Pietro Lio'
F. Precioso
M. Jamnik
G. Marra
NAI
LRM
61
38
0
27 Apr 2023
N$\text{A}^\text{2}$Q: Neural Attention Additive Model for Interpretable
  Multi-Agent Q-Learning
NA2\text{A}^\text{2}A2Q: Neural Attention Additive Model for Interpretable Multi-Agent Q-Learning
Zichuan Liu
Yuanyang Zhu
Chunlin Chen
42
10
0
26 Apr 2023
An Efficient Ensemble Explainable AI (XAI) Approach for Morphed Face
  Detection
An Efficient Ensemble Explainable AI (XAI) Approach for Morphed Face Detection
Rudresh Dwivedi
Ritesh Kumar
Deepak Chopra
Pranay Kothari
Manjot Singh
CVBM
AAML
25
7
0
23 Apr 2023
Explainability in AI Policies: A Critical Review of Communications,
  Reports, Regulations, and Standards in the EU, US, and UK
Explainability in AI Policies: A Critical Review of Communications, Reports, Regulations, and Standards in the EU, US, and UK
L. Nannini
Agathe Balayn
A. Smith
16
37
0
20 Apr 2023
Robustness of Visual Explanations to Common Data Augmentation
Robustness of Visual Explanations to Common Data Augmentation
Lenka Tětková
Lars Kai Hansen
AAML
24
6
0
18 Apr 2023
Evaluating the Robustness of Interpretability Methods through
  Explanation Invariance and Equivariance
Evaluating the Robustness of Interpretability Methods through Explanation Invariance and Equivariance
Jonathan Crabbé
M. Schaar
AAML
24
6
0
13 Apr 2023
Optimizing Data Shapley Interaction Calculation from O(2^n) to O(t n^2)
  for KNN models
Optimizing Data Shapley Interaction Calculation from O(2^n) to O(t n^2) for KNN models
Mohamed Karim Belaid
Dorra El Mekki
Maximilian Rabus
Eyke Hüllermeier
17
3
0
02 Apr 2023
Foundation Models and Fair Use
Foundation Models and Fair Use
Peter Henderson
Xuechen Li
Dan Jurafsky
Tatsunori Hashimoto
Mark A. Lemley
Percy Liang
22
119
0
28 Mar 2023
IDGI: A Framework to Eliminate Explanation Noise from Integrated
  Gradients
IDGI: A Framework to Eliminate Explanation Noise from Integrated Gradients
Ruo Yang
Binghui Wang
M. Bilgic
35
18
0
24 Mar 2023
Revisiting the Fragility of Influence Functions
Revisiting the Fragility of Influence Functions
Jacob R. Epifano
Ravichandran Ramachandran
A. Masino
Ghulam Rasool
TDI
11
14
0
22 Mar 2023
The Representational Status of Deep Learning Models
The Representational Status of Deep Learning Models
Eamon Duede
19
0
0
21 Mar 2023
It Is All About Data: A Survey on the Effects of Data on Adversarial
  Robustness
It Is All About Data: A Survey on the Effects of Data on Adversarial Robustness
Peiyu Xiong
Michael W. Tegegn
Jaskeerat Singh Sarin
Shubhraneel Pal
Julia Rubin
SILM
AAML
32
8
0
17 Mar 2023
A Practical Upper Bound for the Worst-Case Attribution Deviations
A Practical Upper Bound for the Worst-Case Attribution Deviations
Fan Wang
A. Kong
AAML
44
4
0
01 Mar 2023
SUNY: A Visual Interpretation Framework for Convolutional Neural
  Networks from a Necessary and Sufficient Perspective
SUNY: A Visual Interpretation Framework for Convolutional Neural Networks from a Necessary and Sufficient Perspective
Xiwei Xuan
Ziquan Deng
Hsuan-Tien Lin
Z. Kong
Kwan-Liu Ma
AAML
FAtt
24
2
0
01 Mar 2023
Don't be fooled: label leakage in explanation methods and the importance
  of their quantitative evaluation
Don't be fooled: label leakage in explanation methods and the importance of their quantitative evaluation
N. Jethani
A. Saporta
Rajesh Ranganath
FAtt
29
10
0
24 Feb 2023
SplineCam: Exact Visualization and Characterization of Deep Network
  Geometry and Decision Boundaries
SplineCam: Exact Visualization and Characterization of Deep Network Geometry and Decision Boundaries
Ahmed Imtiaz Humayun
Randall Balestriero
Guha Balakrishnan
Richard Baraniuk
24
17
0
24 Feb 2023
The Generalizability of Explanations
The Generalizability of Explanations
Hanxiao Tan
FAtt
15
1
0
23 Feb 2023
Previous
123456...8910
Next