Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2005.01831
Cited By
Evaluating Explainable AI: Which Algorithmic Explanations Help Users Predict Model Behavior?
4 May 2020
Peter Hase
Mohit Bansal
FAtt
Re-assign community
ArXiv
PDF
HTML
Papers citing
"Evaluating Explainable AI: Which Algorithmic Explanations Help Users Predict Model Behavior?"
50 / 62 papers shown
Title
Reasoning Models Don't Always Say What They Think
Yanda Chen
Joe Benton
Ansh Radhakrishnan
Jonathan Uesato
Carson E. Denison
...
Vlad Mikulik
Samuel R. Bowman
Jan Leike
Jared Kaplan
E. Perez
ReLM
LRM
68
12
1
08 May 2025
Exploring the Impact of Explainable AI and Cognitive Capabilities on Users' Decisions
Federico Maria Cau
Lucio Davide Spano
31
0
0
02 May 2025
Gradient Attention Map Based Verification of Deep Convolutional Neural Networks with Application to X-ray Image Datasets
Omid Halimi Milani
Amanda Nikho
Lauren Mills
M. Tliba
Ahmet Enis Cetin
Mohammed H. Elnagar
MedIm
38
0
0
29 Apr 2025
The Impact and Feasibility of Self-Confidence Shaping for AI-Assisted Decision-Making
Takehiro Takayanagi
Ryuji Hashimoto
Chung-Chi Chen
Kiyoshi Izumi
52
0
0
21 Feb 2025
Universal Sparse Autoencoders: Interpretable Cross-Model Concept Alignment
Harrish Thasarathan
Julian Forsyth
Thomas Fel
M. Kowal
Konstantinos G. Derpanis
111
7
0
06 Feb 2025
From Critique to Clarity: A Pathway to Faithful and Personalized Code Explanations with Large Language Models
Zexing Xu
Zhuang Luo
Yichuan Li
Kyumin Lee
S. Rasoul Etesami
38
0
0
28 Jan 2025
"Are You Really Sure?" Understanding the Effects of Human Self-Confidence Calibration in AI-Assisted Decision Making
Shuai Ma
Xinru Wang
Ying Lei
Chuhan Shi
Ming Yin
Xiaojuan Ma
29
24
0
14 Mar 2024
Can Interpretability Layouts Influence Human Perception of Offensive Sentences?
Thiago Freitas dos Santos
Nardine Osman
Marco Schorlemmer
21
0
0
01 Mar 2024
Pyreal: A Framework for Interpretable ML Explanations
Alexandra Zytek
Wei-En Wang
Dongyu Liu
Laure Berti-Equille
K. Veeramachaneni
LRM
37
0
0
20 Dec 2023
ALMANACS: A Simulatability Benchmark for Language Model Explainability
Edmund Mills
Shiye Su
Stuart J. Russell
Scott Emmons
48
7
0
20 Dec 2023
I-CEE: Tailoring Explanations of Image Classification Models to User Expertise
Yao Rong
Peizhu Qian
Vaibhav Unhelkar
Enkelejda Kasneci
34
0
0
19 Dec 2023
Explained anomaly detection in text reviews: Can subjective scenarios be correctly evaluated?
David Novoa-Paradela
O. Fontenla-Romero
B. Guijarro-Berdiñas
20
0
0
08 Nov 2023
InterroLang: Exploring NLP Models and Datasets through Dialogue-based Explanations
Nils Feldhus
Qianli Wang
Tatiana Anikina
Sahil Chopra
Cennet Oguz
Sebastian Möller
32
9
0
09 Oct 2023
Predictability and Comprehensibility in Post-Hoc XAI Methods: A User-Centered Analysis
Anahid N. Jalali
Bernhard Haslhofer
Simone Kriglstein
Andreas Rauber
FAtt
34
4
0
21 Sep 2023
Do Models Explain Themselves? Counterfactual Simulatability of Natural Language Explanations
Yanda Chen
Ruiqi Zhong
Narutatsu Ri
Chen Zhao
He He
Jacob Steinhardt
Zhou Yu
Kathleen McKeown
LRM
26
47
0
17 Jul 2023
MaNtLE: Model-agnostic Natural Language Explainer
Rakesh R Menon
Kerem Zaman
Shashank Srivastava
FAtt
LRM
16
2
0
22 May 2023
Explainability in AI Policies: A Critical Review of Communications, Reports, Regulations, and Standards in the EU, US, and UK
L. Nannini
Agathe Balayn
A. Smith
19
37
0
20 Apr 2023
Understanding the Role of Human Intuition on Reliance in Human-AI Decision-Making with Explanations
Valerie Chen
Q. V. Liao
Jennifer Wortman Vaughan
Gagan Bansal
38
104
0
18 Jan 2023
The State of Human-centered NLP Technology for Fact-checking
Anubrata Das
Houjiang Liu
Venelin Kovatchev
Matthew Lease
HILM
19
61
0
08 Jan 2023
CRAFT: Concept Recursive Activation FacTorization for Explainability
Thomas Fel
Agustin Picard
Louis Bethune
Thibaut Boissin
David Vigouroux
Julien Colin
Rémi Cadène
Thomas Serre
19
102
0
17 Nov 2022
Towards Human-Centred Explainability Benchmarks For Text Classification
Viktor Schlegel
Erick Mendez Guzman
R. Batista-Navarro
18
5
0
10 Nov 2022
Natural Language Deduction with Incomplete Information
Zayne Sprague
Kaj Bostrom
Swarat Chaudhuri
Greg Durrett
LRM
46
17
0
01 Nov 2022
Summarization Programs: Interpretable Abstractive Summarization with Neural Modular Trees
Swarnadeep Saha
Shiyue Zhang
Peter Hase
Mohit Bansal
26
19
0
21 Sep 2022
Beware the Rationalization Trap! When Language Model Explainability Diverges from our Mental Models of Language
R. Sevastjanova
Mennatallah El-Assady
LRM
32
9
0
14 Jul 2022
FRAME: Evaluating Rationale-Label Consistency Metrics for Free-Text Rationales
Aaron Chan
Shaoliang Nie
Liang Tan
Xiaochang Peng
Hamed Firooz
Maziar Sanjabi
Xiang Ren
40
9
0
02 Jul 2022
Towards ML Methods for Biodiversity: A Novel Wild Bee Dataset and Evaluations of XAI Methods for ML-Assisted Rare Species Annotations
Teodor Chiaburu
F. Biessmann
Frank Haußer
32
2
0
15 Jun 2022
Mediators: Conversational Agents Explaining NLP Model Behavior
Nils Feldhus
A. Ravichandran
Sebastian Möller
30
16
0
13 Jun 2022
Use-Case-Grounded Simulations for Explanation Evaluation
Valerie Chen
Nari Johnson
Nicholay Topin
Gregory Plumb
Ameet Talwalkar
FAtt
ELM
22
24
0
05 Jun 2022
GlanceNets: Interpretabile, Leak-proof Concept-based Models
Emanuele Marconato
Andrea Passerini
Stefano Teso
106
64
0
31 May 2022
A Meta-Analysis of the Utility of Explainable Artificial Intelligence in Human-AI Decision-Making
Max Schemmer
Patrick Hemmer
Maximilian Nitsche
Niklas Kühl
Michael Vossing
19
55
0
10 May 2022
ExSum: From Local Explanations to Model Understanding
Yilun Zhou
Marco Tulio Ribeiro
J. Shah
FAtt
LRM
11
25
0
30 Apr 2022
Learning to Scaffold: Optimizing Model Explanations for Teaching
Patrick Fernandes
Marcos Vinícius Treviso
Danish Pruthi
André F. T. Martins
Graham Neubig
FAtt
19
22
0
22 Apr 2022
Perception Visualization: Seeing Through the Eyes of a DNN
Loris Giulivi
Mark J. Carman
Giacomo Boracchi
18
6
0
21 Apr 2022
Calibrating Trust of Multi-Hop Question Answering Systems with Decompositional Probes
Kaige Xie
Sarah Wiegreffe
Mark O. Riedl
ReLM
21
12
0
16 Apr 2022
ProtoTEx: Explaining Model Decisions with Prototype Tensors
Anubrata Das
Chitrank Gupta
Venelin Kovatchev
Matthew Lease
J. Li
24
26
0
11 Apr 2022
Interpreting Language Models with Contrastive Explanations
Kayo Yin
Graham Neubig
MILM
21
77
0
21 Feb 2022
Explainable Predictive Process Monitoring: A User Evaluation
Williams Rizzi
M. Comuzzi
Chiara Di Francescomarino
Chiara Ghidini
Suhwan Lee
F. Maggi
Alexander Nolte
FaML
XAI
16
8
0
15 Feb 2022
Diagnosing AI Explanation Methods with Folk Concepts of Behavior
Alon Jacovi
Jasmijn Bastings
Sebastian Gehrmann
Yoav Goldberg
Katja Filippova
36
15
0
27 Jan 2022
Natural Language Deduction through Search over Statement Compositions
Kaj Bostrom
Zayne Sprague
Swarat Chaudhuri
Greg Durrett
ReLM
LRM
27
46
0
16 Jan 2022
Explain, Edit, and Understand: Rethinking User Study Design for Evaluating Model Explanations
Siddhant Arora
Danish Pruthi
Norman M. Sadeh
William W. Cohen
Zachary Chase Lipton
Graham Neubig
FAtt
35
38
0
17 Dec 2021
UNIREX: A Unified Learning Framework for Language Model Rationale Extraction
Aaron Chan
Maziar Sanjabi
Lambert Mathias
L Tan
Shaoliang Nie
Xiaochang Peng
Xiang Ren
Hamed Firooz
41
41
0
16 Dec 2021
HIVE: Evaluating the Human Interpretability of Visual Explanations
Sunnie S. Y. Kim
Nicole Meister
V. V. Ramaswamy
Ruth C. Fong
Olga Russakovsky
66
114
0
06 Dec 2021
Evaluating the Faithfulness of Importance Measures in NLP by Recursively Masking Allegedly Important Tokens and Retraining
Andreas Madsen
Nicholas Meade
Vaibhav Adlakha
Siva Reddy
103
35
0
15 Oct 2021
Can Explanations Be Useful for Calibrating Black Box Models?
Xi Ye
Greg Durrett
FAtt
24
25
0
14 Oct 2021
Counterfactual Evaluation for Explainable AI
Yingqiang Ge
Shuchang Liu
Zelong Li
Shuyuan Xu
Shijie Geng
Yunqi Li
Juntao Tan
Fei Sun
Yongfeng Zhang
CML
35
13
0
05 Sep 2021
How Well do Feature Visualizations Support Causal Understanding of CNN Activations?
Roland S. Zimmermann
Judy Borowski
Robert Geirhos
Matthias Bethge
Thomas S. A. Wallis
Wieland Brendel
FAtt
39
31
0
23 Jun 2021
multiPRover: Generating Multiple Proofs for Improved Interpretability in Rule Reasoning
Swarnadeep Saha
Prateek Yadav
Mohit Bansal
ReLM
LRM
16
26
0
02 Jun 2021
The Out-of-Distribution Problem in Explainability and Search Methods for Feature Importance Explanations
Peter Hase
Harry Xie
Mohit Bansal
OODD
LRM
FAtt
18
91
0
01 Jun 2021
Explainable Artificial Intelligence for Human Decision-Support System in Medical Domain
Samanta Knapic
A. Malhi
Rohit Saluja
Kary Främling
13
99
0
05 May 2021
On the Sensitivity and Stability of Model Interpretations in NLP
Fan Yin
Zhouxing Shi
Cho-Jui Hsieh
Kai-Wei Chang
FAtt
11
33
0
18 Apr 2021
1
2
Next