ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2202.00734
  4. Cited By
Framework for Evaluating Faithfulness of Local Explanations

Framework for Evaluating Faithfulness of Local Explanations

International Conference on Machine Learning (ICML), 2022
1 February 2022
S. Dasgupta
Nave Frost
Michal Moshkovitz
    FAtt
ArXiv (abs)PDFHTMLGithub

Papers citing "Framework for Evaluating Faithfulness of Local Explanations"

44 / 44 papers shown
Additive Large Language Models for Semi-Structured Text
Additive Large Language Models for Semi-Structured Text
Karthikeyan K
Raghuveer Thirukovalluru
David Edwin Carlson
152
0
0
14 Nov 2025
Faithful-First Reasoning, Planning, and Acting for Multimodal LLMs
Faithful-First Reasoning, Planning, and Acting for Multimodal LLMs
Junxian Li
Xinyue Xu
Sai Ma
Sichao Li
Sichao Li
LRM
230
1
0
11 Nov 2025
rCamInspector: Building Reliability and Trust on IoT (Spy) Camera Detection using XAI
rCamInspector: Building Reliability and Trust on IoT (Spy) Camera Detection using XAI
P. Chaudhary
Manan Gupta
Jabez Christopher
Putrevu Venkata Sai Charan
R. Maiti
162
0
0
12 Sep 2025
Value bounds and Convergence Analysis for Averages of LRP attributions
Value bounds and Convergence Analysis for Averages of LRP attributions
Alexander Binder
Nastaran Takmil-Homayouni
Ürün Dogan
FAtt
298
0
0
10 Sep 2025
Assessing the Noise Robustness of Class Activation Maps: A Framework for Reliable Model Interpretability
Assessing the Noise Robustness of Class Activation Maps: A Framework for Reliable Model InterpretabilityImage and Vision Computing (IVC), 2025
Syamantak Sarkar
Revoti P. Bora
Bhupender Kaushal
Sudhish N George
Kiran Raja
AAML
125
0
0
25 Aug 2025
Informative Post-Hoc Explanations Only Exist for Simple Functions
Informative Post-Hoc Explanations Only Exist for Simple Functions
Eric Günther
Balázs Szabados
Robi Bhattacharjee
Sebastian Bordt
U. V. Luxburg
FAtt
232
3
0
15 Aug 2025
DeepFaith: A Domain-Free and Model-Agnostic Unified Framework for Highly Faithful Explanations
DeepFaith: A Domain-Free and Model-Agnostic Unified Framework for Highly Faithful Explanations
Yuhan Guo
Lizhong Ding
Shihan Jia
Yanyu Ren
P. Li
Jiarun Fu
Changsheng Li
Ye Yuan
Guoren Wang
230
0
0
05 Aug 2025
Argument-Based Consistency in Toxicity Explanations of LLMs
Argument-Based Consistency in Toxicity Explanations of LLMs
Ramaravind Kommiya Mothilal
Joanna Roy
Syed Ishtiaque Ahmed
Shion Guha
233
0
0
23 Jun 2025
Differentially Private Explanations for Clusters
Differentially Private Explanations for Clusters
Amir Gilad
Tova Milo
Kathy Razmadze
Ron Zadicario
233
1
0
06 Jun 2025
Privacy Risks and Preservation Methods in Explainable Artificial Intelligence: A Scoping Review
Privacy Risks and Preservation Methods in Explainable Artificial Intelligence: A Scoping Review
Sonal Allana
Mohan Kankanhalli
Rozita Dara
442
6
0
05 May 2025
Explanations Go Linear: Post-hoc Explainability for Tabular Data with Interpretable Meta-Encoding
Explanations Go Linear: Post-hoc Explainability for Tabular Data with Interpretable Meta-Encoding
Simone Piaggesi
Riccardo Guidotti
F. Giannotti
D. Pedreschi
FAttMILMLRM
1.2K
0
0
29 Apr 2025
Beyond Patches: Mining Interpretable Part-Prototypes for Explainable AI
Beyond Patches: Mining Interpretable Part-Prototypes for Explainable AI
Mahdi Alehdaghi
Rajarshi Bhattacharya
Pourya Shamsolmoali
Rafael M. O. Cruz
Maguelonne Heritier
Mohammadhadi Shateri
501
1
0
16 Apr 2025
Towards an Evaluation Framework for Explainable Artificial Intelligence Systems for Health and Well-being
Towards an Evaluation Framework for Explainable Artificial Intelligence Systems for Health and Well-beingInternational Conference on Evaluation of Novel Approaches to Software Engineering (ENASE), 2025
Esperança Amengual-Alcover
Antoni Jaume-i-Capó
Miquel Miró-Nicolau
Gabriel Moyà Alcover
Antonia Paniza-Fullana
304
2
0
11 Apr 2025
How to safely discard features based on aggregate SHAP values
How to safely discard features based on aggregate SHAP valuesAnnual Conference Computational Learning Theory (COLT), 2025
Robi Bhattacharjee
Karolin Frohnapfel
U. V. Luxburg
TDIFAtt
426
4
0
29 Mar 2025
Evaluate with the Inverse: Efficient Approximation of Latent Explanation Quality Distribution
Evaluate with the Inverse: Efficient Approximation of Latent Explanation Quality DistributionAAAI Conference on Artificial Intelligence (AAAI), 2025
Carlos Eiras-Franco
Anna Hedström
Marina M.-C. Höhne
XAI
276
0
0
24 Feb 2025
ExpProof : Operationalizing Explanations for Confidential Models with ZKPs
ExpProof : Operationalizing Explanations for Confidential Models with ZKPs
Chhavi Yadav
Evan Monroe Laufer
Dan Boneh
Kamalika Chaudhuri
577
3
0
06 Feb 2025
Navigating the Maze of Explainable AI: A Systematic Approach to Evaluating Methods and Metrics
Navigating the Maze of Explainable AI: A Systematic Approach to Evaluating Methods and MetricsNeural Information Processing Systems (NeurIPS), 2024
Lukas Klein
Carsten T. Lüth
U. Schlegel
Till J. Bungert
Mennatallah El-Assady
Paul F. Jäger
XAIELM
743
22
0
03 Jan 2025
Advancing Attribution-Based Neural Network Explainability through
  Relative Absolute Magnitude Layer-Wise Relevance Propagation and
  Multi-Component Evaluation
Advancing Attribution-Based Neural Network Explainability through Relative Absolute Magnitude Layer-Wise Relevance Propagation and Multi-Component EvaluationACM Transactions on Intelligent Systems and Technology (ACM TIST), 2024
Davor Vukadin
Petar Afrić
Marin Šilić
Goran Delač
FAtt
306
2
0
12 Dec 2024
From Flexibility to Manipulation: The Slippery Slope of XAI Evaluation
From Flexibility to Manipulation: The Slippery Slope of XAI Evaluation
Kristoffer Wickstrøm
Marina M.-C. Höhne
Anna Hedström
AAML
481
6
0
07 Dec 2024
GIFT: A Framework Towards Global Interpretable Faithful Textual Explanations of Vision Classifiers
GIFT: A Framework Towards Global Interpretable Faithful Textual Explanations of Vision Classifiers
Éloi Zablocki
Valentin Gerard
Amaia Cardiel
Eric Gaussier
Matthieu Cord
Eduardo Valle
545
0
0
23 Nov 2024
Benchmarking XAI Explanations with Human-Aligned Evaluations
Benchmarking XAI Explanations with Human-Aligned Evaluations
Rémi Kazmierczak
Steve Azzolin
Eloise Berthier
Anna Hedström
Patricia Delhomme
...
Goran Frehse
Baptiste Caramiaux
Baptiste Caramiaux
Andrea Passerini
Gianni Franchi
545
7
0
04 Nov 2024
IBO: Inpainting-Based Occlusion to Enhance Explainable Artificial
  Intelligence Evaluation in Histopathology
IBO: Inpainting-Based Occlusion to Enhance Explainable Artificial Intelligence Evaluation in Histopathology
Pardis Afshar
Sajjad Hashembeiki
Pouya Khani
Emad Fatemizadeh
M. Rohban
396
5
0
29 Aug 2024
Auditing Local Explanations is Hard
Auditing Local Explanations is Hard
Robi Bhattacharjee
U. V. Luxburg
LRMMLAUFAtt
331
8
0
18 Jul 2024
Expected Grad-CAM: Towards gradient faithfulness
Expected Grad-CAM: Towards gradient faithfulness
Vincenzo Buono
Peyman Sheikholharam Mashhadi
M. Rahat
Prayag Tiwari
Stefan Byttner
FAtt
312
4
0
03 Jun 2024
DISCRET: Synthesizing Faithful Explanations For Treatment Effect
  Estimation
DISCRET: Synthesizing Faithful Explanations For Treatment Effect Estimation
Yinjun Wu
Mayank Keoliya
Kan Chen
Neelay Velingker
Ziyang Li
E. Getzen
Qi Long
Mayur Naik
Ravi B. Parikh
Eric Wong
324
3
0
02 Jun 2024
A Sim2Real Approach for Identifying Task-Relevant Properties in
  Interpretable Machine Learning
A Sim2Real Approach for Identifying Task-Relevant Properties in Interpretable Machine Learning
Eura Nofshin
Esther Brown
Brian Lim
Weiwei Pan
Finale Doshi-Velez
392
1
0
31 May 2024
Towards a Novel Measure of User Trust in XAI Systems
Towards a Novel Measure of User Trust in XAI Systems
Miquel Miró-Nicolau
Gabriel Moyà Alcover
Antoni Jaume-i-Capó
Manuel González Hidalgo
Adel Ghazel
Maria Gemma Sempere Campello
Juan Antonio Palmer Sancho
348
0
0
09 May 2024
A Fresh Look at Sanity Checks for Saliency Maps
A Fresh Look at Sanity Checks for Saliency Maps
Anna Hedström
Leander Weber
Sebastian Lapuschkin
Marina M.-C. Höhne
FAttLRM
352
16
0
03 May 2024
Toward Understanding the Disagreement Problem in Neural Network Feature
  Attribution
Toward Understanding the Disagreement Problem in Neural Network Feature Attribution
Niklas Koenen
Marvin N. Wright
FAtt
246
13
0
17 Apr 2024
Sanity Checks Revisited: An Exploration to Repair the Model Parameter
  Randomisation Test
Sanity Checks Revisited: An Exploration to Repair the Model Parameter Randomisation Test
Anna Hedström
Leander Weber
Sebastian Lapuschkin
Marina M.-C. Höhne
LRM
350
10
0
12 Jan 2024
Proto-lm: A Prototypical Network-Based Framework for Built-in
  Interpretability in Large Language Models
Proto-lm: A Prototypical Network-Based Framework for Built-in Interpretability in Large Language ModelsConference on Empirical Methods in Natural Language Processing (EMNLP), 2023
Sean Xie
Soroush Vosoughi
Saeed Hassanpour
373
8
0
03 Nov 2023
Faithful and Robust Local Interpretability for Textual Predictions
Faithful and Robust Local Interpretability for Textual Predictions
Gianluigi Lopardo
F. Precioso
Damien Garreau
OOD
347
4
0
30 Oct 2023
A Framework for Interpretability in Machine Learning for Medical Imaging
A Framework for Interpretability in Machine Learning for Medical ImagingIEEE Access (IEEE Access), 2023
Alan Q. Wang
Batuhan K. Karaman
Heejong Kim
Jacob Rosenthal
Rachit Saluja
Sean I. Young
M. Sabuncu
AI4CE
524
29
0
02 Oct 2023
Faithful Explanations of Black-box NLP Models Using LLM-generated
  Counterfactuals
Faithful Explanations of Black-box NLP Models Using LLM-generated CounterfactualsInternational Conference on Learning Representations (ICLR), 2023
Y. Gat
Nitay Calderon
Amir Feder
Alexander Chapanin
Amit Sharma
Roi Reichart
438
51
0
01 Oct 2023
Explainability for Large Language Models: A Survey
Explainability for Large Language Models: A SurveyACM Transactions on Intelligent Systems and Technology (ACM TIST), 2023
Haiyan Zhao
Hanjie Chen
Fan Yang
Ninghao Liu
Huiqi Deng
Hengyi Cai
Shuaiqiang Wang
D. Yin
Jundong Li
LRM
584
793
0
02 Sep 2023
Encoding Time-Series Explanations through Self-Supervised Model Behavior
  Consistency
Encoding Time-Series Explanations through Self-Supervised Model Behavior ConsistencyNeural Information Processing Systems (NeurIPS), 2023
Owen Queen
Thomas Hartvigsen
Teddy Koker
Huan He
Theodoros Tsiligkaridis
Marinka Zitnik
AI4TS
396
40
0
03 Jun 2023
Feature Perturbation Augmentation for Reliable Evaluation of Importance
  Estimators in Neural Networks
Feature Perturbation Augmentation for Reliable Evaluation of Importance Estimators in Neural NetworksPattern Recognition Letters (PR), 2023
L. Brocki
N. C. Chung
FAttAAML
337
21
0
02 Mar 2023
The Meta-Evaluation Problem in Explainable AI: Identifying Reliable
  Estimators with MetaQuantus
The Meta-Evaluation Problem in Explainable AI: Identifying Reliable Estimators with MetaQuantus
Anna Hedström
P. Bommer
Kristoffer K. Wickstrom
Wojciech Samek
Sebastian Lapuschkin
Marina M.-C. Höhne
364
35
0
14 Feb 2023
A Survey of Explainable AI in Deep Visual Modeling: Methods and Metrics
A Survey of Explainable AI in Deep Visual Modeling: Methods and Metrics
Naveed Akhtar
XAIVLM
273
9
0
31 Jan 2023
SAFARI: Versatile and Efficient Evaluations for Robustness of
  Interpretability
SAFARI: Versatile and Efficient Evaluations for Robustness of InterpretabilityIEEE International Conference on Computer Vision (ICCV), 2022
Wei Huang
Xingyu Zhao
Gao Jin
Xiaowei Huang
AAML
476
40
0
19 Aug 2022
OpenXAI: Towards a Transparent Evaluation of Model Explanations
OpenXAI: Towards a Transparent Evaluation of Model Explanations
Chirag Agarwal
Dan Ley
Satyapriya Krishna
Eshika Saxena
Martin Pawelczyk
Nari Johnson
Isha Puri
Marinka Zitnik
Himabindu Lakkaraju
XAI
617
190
0
22 Jun 2022
XAudit : A Theoretical Look at Auditing with Explanations
XAudit : A Theoretical Look at Auditing with Explanations
Chhavi Yadav
Michal Moshkovitz
Kamalika Chaudhuri
XAIFAttMLAU
369
5
0
09 Jun 2022
Quantus: An Explainable AI Toolkit for Responsible Evaluation of Neural
  Network Explanations and Beyond
Quantus: An Explainable AI Toolkit for Responsible Evaluation of Neural Network Explanations and BeyondJournal of machine learning research (JMLR), 2022
Anna Hedström
Leander Weber
Dilyara Bareeva
Daniel G. Krakowczyk
Franz Motzkus
Wojciech Samek
Sebastian Lapuschkin
Marina M.-C. Höhne
XAIELM
444
245
0
14 Feb 2022
"Why Should I Trust You?": Explaining the Predictions of Any Classifier
"Why Should I Trust You?": Explaining the Predictions of Any Classifier
Marco Tulio Ribeiro
Sameer Singh
Carlos Guestrin
FAttFaML
2.7K
21,148
0
16 Feb 2016
1
Page 1 of 1