ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2106.01087
  4. Cited By
Is Sparse Attention more Interpretable?
v1v2 (latest)

Is Sparse Attention more Interpretable?

Annual Meeting of the Association for Computational Linguistics (ACL), 2021
2 June 2021
Clara Meister
Stefan Lazov
Isabelle Augenstein
Robert Bamler
    MILM
ArXiv (abs)PDFHTML

Papers citing "Is Sparse Attention more Interpretable?"

29 / 29 papers shown
Title
Understanding Sensitivity of Differential Attention through the Lens of Adversarial Robustness
Understanding Sensitivity of Differential Attention through the Lens of Adversarial Robustness
Tsubasa Takahashi
Shojiro Yamabe
Futa Waseda
Kento Sasaki
AAML
92
0
0
01 Oct 2025
Hierarchical Attention Network for Interpretable ECG-based Heart Disease Classification
Hierarchical Attention Network for Interpretable ECG-based Heart Disease Classification
Mario Padilla Rodriguez
Mohamed Nafea
142
3
0
25 Mar 2025
Membership Inference Attack against Long-Context Large Language Models
Zixiong Wang
Gaoyang Liu
Yang Yang
Chen Wang
318
3
0
18 Nov 2024
$k$NN Attention Demystified: A Theoretical Exploration for Scalable
  Transformers
kkkNN Attention Demystified: A Theoretical Exploration for Scalable Transformers
Themistoklis Haris
230
0
0
06 Nov 2024
Probing Ranking LLMs: A Mechanistic Analysis for Information Retrieval
Probing Ranking LLMs: A Mechanistic Analysis for Information RetrievalInternational Conference on the Theory of Information Retrieval (ICTIR), 2024
Tanya Chowdhury
Atharva Nijasure
James Allan
183
2
0
24 Oct 2024
Mechanistic?
Mechanistic?BlackboxNLP Workshop on Analyzing and Interpreting Neural Networks for NLP (BlackBoxNLP), 2024
Naomi Saphra
Sarah Wiegreffe
AI4CE
201
31
0
07 Oct 2024
DreamStory: Open-Domain Story Visualization by LLM-Guided Multi-Subject Consistent Diffusion
DreamStory: Open-Domain Story Visualization by LLM-Guided Multi-Subject Consistent Diffusion
Huiguo He
Huan Yang
Zixi Tuo
Yuan Zhou
Qiuyue Wang
Yuhang Zhang
Zeyu Liu
Wenhao Huang
Hongyang Chao
Jian Yin
VGenDiffM
431
27
0
17 Jul 2024
Interpretability Needs a New Paradigm
Interpretability Needs a New Paradigm
Andreas Madsen
Himabindu Lakkaraju
Siva Reddy
Sarath Chandar
177
6
0
08 May 2024
Towards Explainability in Legal Outcome Prediction Models
Towards Explainability in Legal Outcome Prediction Models
Josef Valvoda
Robert Bamler
ELMAILaw
227
5
0
25 Mar 2024
Sequence Shortening for Context-Aware Machine Translation
Sequence Shortening for Context-Aware Machine Translation
Paweł Mąka
Yusuf Can Semerci
Jan Scholtes
Gerasimos Spanakis
122
3
0
02 Feb 2024
Towards Faithful Explanations for Text Classification with Robustness
  Improvement and Explanation Guided Training
Towards Faithful Explanations for Text Classification with Robustness Improvement and Explanation Guided Training
Dongfang Li
Baotian Hu
Qingcai Chen
Shan He
235
7
0
29 Dec 2023
Sparsity-Guided Holistic Explanation for LLMs with Interpretable
  Inference-Time Intervention
Sparsity-Guided Holistic Explanation for LLMs with Interpretable Inference-Time Intervention
Zhen Tan
Tianlong Chen
Zhenyu Zhang
Huan Liu
152
24
0
22 Dec 2023
Transformers are uninterpretable with myopic methods: a case study with
  bounded Dyck grammars
Transformers are uninterpretable with myopic methods: a case study with bounded Dyck grammarsNeural Information Processing Systems (NeurIPS), 2023
Kaiyue Wen
Yuchen Li
Bing Liu
Andrej Risteski
238
27
0
03 Dec 2023
LooGLE: Can Long-Context Language Models Understand Long Contexts?
LooGLE: Can Long-Context Language Models Understand Long Contexts?
Jiaqi Li
Minghua Yi
Zilong Zheng
Muhan Zhang
ELMRALM
232
180
0
08 Nov 2023
Faithfulness Measurable Masked Language Models
Faithfulness Measurable Masked Language ModelsInternational Conference on Machine Learning (ICML), 2023
Andreas Madsen
Siva Reddy
Sarath Chandar
174
4
0
11 Oct 2023
Sparse Autoencoders Find Highly Interpretable Features in Language
  Models
Sparse Autoencoders Find Highly Interpretable Features in Language ModelsInternational Conference on Learning Representations (ICLR), 2023
Hoagy Cunningham
Aidan Ewart
Logan Riggs
R. Huben
Lee Sharkey
MILM
518
745
0
15 Sep 2023
Exposing Attention Glitches with Flip-Flop Language Modeling
Exposing Attention Glitches with Flip-Flop Language ModelingNeural Information Processing Systems (NeurIPS), 2023
Bingbin Liu
Jordan T. Ash
Surbhi Goel
A. Krishnamurthy
Cyril Zhang
LRM
185
70
0
01 Jun 2023
Consistent Multi-Granular Rationale Extraction for Explainable Multi-hop
  Fact Verification
Consistent Multi-Granular Rationale Extraction for Explainable Multi-hop Fact Verification
Jiasheng Si
Yingjie Zhu
Deyu Zhou
AAML
277
5
0
16 May 2023
Explainability of Text Processing and Retrieval Methods: A Survey
Explainability of Text Processing and Retrieval Methods: A Survey
Sourav Saha
Debapriyo Majumdar
Mandar Mitra
254
5
0
14 Dec 2022
Exploring Faithful Rationale for Multi-hop Fact Verification via
  Salience-Aware Graph Learning
Exploring Faithful Rationale for Multi-hop Fact Verification via Salience-Aware Graph LearningAAAI Conference on Artificial Intelligence (AAAI), 2022
Jiasheng Si
Yingjie Zhu
Deyu Zhou
246
22
0
02 Dec 2022
Easy to Decide, Hard to Agree: Reducing Disagreements Between Saliency
  Methods
Easy to Decide, Hard to Agree: Reducing Disagreements Between Saliency MethodsAnnual Meeting of the Association for Computational Linguistics (ACL), 2022
Josip Jukić
Martin Tutek
Jan Snajder
FAtt
213
0
0
15 Nov 2022
Exploring Self-Attention for Crop-type Classification Explainability
Exploring Self-Attention for Crop-type Classification Explainability
Ivica Obadic
R. Roscher
Dario Augusto Borges Oliveira
Xiao Xiang Zhu
274
7
0
24 Oct 2022
Breaking BERT: Evaluating and Optimizing Sparsified Attention
Breaking BERT: Evaluating and Optimizing Sparsified Attention
Siddhartha Brahma
Polina Zablotskaia
David M. Mimno
114
1
0
07 Oct 2022
Toward Transparent AI: A Survey on Interpreting the Inner Structures of
  Deep Neural Networks
Toward Transparent AI: A Survey on Interpreting the Inner Structures of Deep Neural Networks
Tilman Raukur
A. Ho
Stephen Casper
Dylan Hadfield-Menell
AAMLAI4CE
678
164
0
27 Jul 2022
A Song of (Dis)agreement: Evaluating the Evaluation of Explainable
  Artificial Intelligence in Natural Language Processing
A Song of (Dis)agreement: Evaluating the Evaluation of Explainable Artificial Intelligence in Natural Language Processing
Michael Neely
Stefan F. Schouten
Maurits J. R. Bleeker
Ana Lucic
XAI
178
19
0
09 May 2022
Evaluating the Faithfulness of Importance Measures in NLP by Recursively
  Masking Allegedly Important Tokens and Retraining
Evaluating the Faithfulness of Importance Measures in NLP by Recursively Masking Allegedly Important Tokens and Retraining
Andreas Madsen
Nicholas Meade
Vaibhav Adlakha
Siva Reddy
269
41
0
15 Oct 2021
How Does Adversarial Fine-Tuning Benefit BERT?
How Does Adversarial Fine-Tuning Benefit BERT?
J. Ebrahimi
Hao Yang
Wei Zhang
AAML
169
5
0
31 Aug 2021
Post-hoc Interpretability for Neural NLP: A Survey
Post-hoc Interpretability for Neural NLP: A SurveyACM Computing Surveys (CSUR), 2021
Andreas Madsen
Siva Reddy
A. Chandar
XAI
281
269
0
10 Aug 2021
Making Attention Mechanisms More Robust and Interpretable with Virtual
  Adversarial Training
Making Attention Mechanisms More Robust and Interpretable with Virtual Adversarial Training
Shunsuke Kitada
Hitoshi Iyatomi
AAML
304
10
0
18 Apr 2021
1