ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2504.17023
  4. Cited By
What Makes for a Good Saliency Map? Comparing Strategies for Evaluating Saliency Maps in Explainable AI (XAI)

What Makes for a Good Saliency Map? Comparing Strategies for Evaluating Saliency Maps in Explainable AI (XAI)

23 April 2025
Felix Kares
Timo Speith
Hanwei Zhang
Markus Langer
    FAttXAI
ArXiv (abs)PDFHTMLGithub (12704★)

Papers citing "What Makes for a Good Saliency Map? Comparing Strategies for Evaluating Saliency Maps in Explainable AI (XAI)"

3 / 3 papers shown
See What I Mean? CUE: A Cognitive Model of Understanding Explanations
See What I Mean? CUE: A Cognitive Model of Understanding Explanations
Tobias Labarta
Nhi Hoang
Katharina Weitz
Wojciech Samek
Sebastian Lapuschkin
Leander Weber
289
1
0
09 May 2025
Benchmarking XAI Explanations with Human-Aligned Evaluations
Benchmarking XAI Explanations with Human-Aligned Evaluations
Rémi Kazmierczak
Steve Azzolin
Eloise Berthier
Anna Hedström
Patricia Delhomme
...
Goran Frehse
Baptiste Caramiaux
Baptiste Caramiaux
Andrea Passerini
Gianni Franchi
539
6
0
04 Nov 2024
"Why Should I Trust You?": Explaining the Predictions of Any Classifier
"Why Should I Trust You?": Explaining the Predictions of Any Classifier
Marco Tulio Ribeiro
Sameer Singh
Carlos Guestrin
FAttFaML
2.7K
21,148
0
16 Feb 2016
1
Page 1 of 1