ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2005.00928
  4. Cited By
Quantifying Attention Flow in Transformers

Quantifying Attention Flow in Transformers

2 May 2020
Samira Abnar
Willem H. Zuidema
ArXivPDFHTML

Papers citing "Quantifying Attention Flow in Transformers"

50 / 403 papers shown
Title
Progressive Inference: Explaining Decoder-Only Sequence Classification
  Models Using Intermediate Predictions
Progressive Inference: Explaining Decoder-Only Sequence Classification Models Using Intermediate Predictions
Sanjay Kariyappa
Freddy Lecue
Saumitra Mishra
Christopher Pond
Daniele Magazzeni
Manuela Veloso
43
1
0
03 Jun 2024
MultiMax: Sparse and Multi-Modal Attention Learning
MultiMax: Sparse and Multi-Modal Attention Learning
Yuxuan Zhou
Mario Fritz
M. Keuper
42
1
0
03 Jun 2024
DeCo: Decoupling Token Compression from Semantic Abstraction in
  Multimodal Large Language Models
DeCo: Decoupling Token Compression from Semantic Abstraction in Multimodal Large Language Models
Linli Yao
Lei Li
Shuhuai Ren
Lean Wang
Yuanxin Liu
Xu Sun
Lu Hou
35
29
0
31 May 2024
Scalable Numerical Embeddings for Multivariate Time Series: Enhancing
  Healthcare Data Representation Learning
Scalable Numerical Embeddings for Multivariate Time Series: Enhancing Healthcare Data Representation Learning
Chun-Kai Huang
Yi-Hsien Hsieh
Ta-Jung Chien
Li-Cheng Chien
Shao-Hua Sun
T. Su
J. Kao
Che Lin
AI4TS
45
0
0
26 May 2024
Explainable Molecular Property Prediction: Aligning Chemical Concepts
  with Predictions via Language Models
Explainable Molecular Property Prediction: Aligning Chemical Concepts with Predictions via Language Models
Zhenzhong Wang
Zehui Lin
Wanyu Lin
Ming Yang
Minggang Zeng
Kay Chen Tan
28
3
0
25 May 2024
Recasting Generic Pretrained Vision Transformers As Object-Centric Scene
  Encoders For Manipulation Policies
Recasting Generic Pretrained Vision Transformers As Object-Centric Scene Encoders For Manipulation Policies
Jianing Qian
Anastasios Panagopoulos
Dinesh Jayaraman
LM&Ro
ViT
38
5
0
24 May 2024
Explaining Multi-modal Large Language Models by Analyzing their Vision
  Perception
Explaining Multi-modal Large Language Models by Analyzing their Vision Perception
Loris Giulivi
Giacomo Boracchi
38
2
0
23 May 2024
Empowering Prior to Court Legal Analysis: A Transparent and Accessible
  Dataset for Defensive Statement Classification and Interpretation
Empowering Prior to Court Legal Analysis: A Transparent and Accessible Dataset for Defensive Statement Classification and Interpretation
Yannis Spyridis
Jean-Paul Younes
Haneen Deeb
Vasileios Argyriou
27
0
0
17 May 2024
Divergent Creativity in Humans and Large Language Models
Divergent Creativity in Humans and Large Language Models
Antoine Bellemare-Pepin
Franccois Lespinasse
Philipp Tholke
Y. Harel
K. Mathewson
Jay A. Olson
Yoshua Bengio
Department of Computer Science
AI4CE
45
9
0
13 May 2024
Explaining Text Similarity in Transformer Models
Explaining Text Similarity in Transformer Models
Alexandros Vasileiou
Oliver Eberle
43
7
0
10 May 2024
MR-Transformer: Vision Transformer for Total Knee Replacement Prediction
  Using Magnetic Resonance Imaging
MR-Transformer: Vision Transformer for Total Knee Replacement Prediction Using Magnetic Resonance Imaging
Chaojie Zhang
Shengjia Chen
O. Cigdem
H. Rajamohan
K. Cho
R. Kijowski
Cem M. Deniz
MedIm
ViT
18
1
0
05 May 2024
Made to Order: Discovering monotonic temporal changes via
  self-supervised video ordering
Made to Order: Discovering monotonic temporal changes via self-supervised video ordering
Charig Yang
Weidi Xie
Andrew Zisserman
34
1
0
25 Apr 2024
SNP: Structured Neuron-level Pruning to Preserve Attention Scores
SNP: Structured Neuron-level Pruning to Preserve Attention Scores
Kyunghwan Shim
Jaewoong Yun
Shinkook Choi
30
0
0
18 Apr 2024
Weight Copy and Low-Rank Adaptation for Few-Shot Distillation of Vision
  Transformers
Weight Copy and Low-Rank Adaptation for Few-Shot Distillation of Vision Transformers
Diana-Nicoleta Grigore
Mariana-Iuliana Georgescu
J. A. Justo
T. Johansen
Andreea-Iuliana Ionescu
Radu Tudor Ionescu
36
0
0
14 Apr 2024
How does Multi-Task Training Affect Transformer In-Context Capabilities?
  Investigations with Function Classes
How does Multi-Task Training Affect Transformer In-Context Capabilities? Investigations with Function Classes
Harmon Bhasin
Timothy Ossowski
Yiqiao Zhong
Junjie Hu
24
0
0
04 Apr 2024
AdaGlimpse: Active Visual Exploration with Arbitrary Glimpse Position
  and Scale
AdaGlimpse: Active Visual Exploration with Arbitrary Glimpse Position and Scale
Adam Pardyl
Michal Wronka
Maciej Wolczyk
Kamil Adamczewski
Tomasz Trzciñski
Bartosz Zieliñski
41
2
0
04 Apr 2024
LeGrad: An Explainability Method for Vision Transformers via Feature Formation Sensitivity
LeGrad: An Explainability Method for Vision Transformers via Feature Formation Sensitivity
Walid Bousselham
Angie Boggust
Sofian Chaybouti
Hendrik Strobelt
Hilde Kuehne
96
10
0
04 Apr 2024
DeiT-LT Distillation Strikes Back for Vision Transformer Training on
  Long-Tailed Datasets
DeiT-LT Distillation Strikes Back for Vision Transformer Training on Long-Tailed Datasets
Harsh Rangwani
Pradipto Mondal
Mayank Mishra
Ashish Ramayee Asokan
R. V. Babu
37
8
0
03 Apr 2024
Masked Completion via Structured Diffusion with White-Box Transformers
Masked Completion via Structured Diffusion with White-Box Transformers
Druv Pai
Ziyang Wu
Sam Buchanan
Yaodong Yu
Yi Ma
35
13
0
03 Apr 2024
On the Faithfulness of Vision Transformer Explanations
On the Faithfulness of Vision Transformer Explanations
Junyi Wu
Weitai Kang
Hao Tang
Yuan Hong
Yan Yan
27
6
0
01 Apr 2024
Does Faithfulness Conflict with Plausibility? An Empirical Study in
  Explainable AI across NLP Tasks
Does Faithfulness Conflict with Plausibility? An Empirical Study in Explainable AI across NLP Tasks
Xiaolei Lu
Jianghong Ma
23
0
0
29 Mar 2024
Enhancing Efficiency in Vision Transformer Networks: Design Techniques
  and Insights
Enhancing Efficiency in Vision Transformer Networks: Design Techniques and Insights
Moein Heidari
Reza Azad
Sina Ghorbani Kolahi
René Arimond
Leon Niggemeier
...
Afshin Bozorgpour
Ehsan Khodapanah Aghdam
A. Kazerouni
I. Hacihaliloglu
Dorit Merhof
51
7
0
28 Mar 2024
PE: A Poincare Explanation Method for Fast Text Hierarchy Generation
PE: A Poincare Explanation Method for Fast Text Hierarchy Generation
Qian Chen
Dongyang Li
Xiaofeng He
Hongzhao Li
Hongyu Yi
24
0
0
25 Mar 2024
Token Transformation Matters: Towards Faithful Post-hoc Explanation for
  Vision Transformer
Token Transformation Matters: Towards Faithful Post-hoc Explanation for Vision Transformer
Junyi Wu
Bin Duan
Weitai Kang
Hao Tang
Yan Yan
36
6
0
21 Mar 2024
vid-TLDR: Training Free Token merging for Light-weight Video Transformer
vid-TLDR: Training Free Token merging for Light-weight Video Transformer
Joonmyung Choi
Sanghyeok Lee
Jaewon Chu
Minhyuk Choi
Hyunwoo J. Kim
MoMe
ViT
55
12
0
20 Mar 2024
SelfIE: Self-Interpretation of Large Language Model Embeddings
SelfIE: Self-Interpretation of Large Language Model Embeddings
Haozhe Chen
Carl Vondrick
Chengzhi Mao
27
18
0
16 Mar 2024
Approximate Nullspace Augmented Finetuning for Robust Vision Transformers
Approximate Nullspace Augmented Finetuning for Robust Vision Transformers
Haoyang Liu
Aditya Singh
Yijiang Li
Haohan Wang
AAML
ViT
39
1
0
15 Mar 2024
Matching Non-Identical Objects
Matching Non-Identical Objects
Yusuke Marumo
Kazuhiko Kawamoto
Hiroshi Kera
32
0
0
13 Mar 2024
IndicSTR12: A Dataset for Indic Scene Text Recognition
IndicSTR12: A Dataset for Indic Scene Text Recognition
Harsh Lunia
Ajoy Mondal
C. V. Jawahar
24
2
0
12 Mar 2024
Robust Wake Word Spotting With Frame-Level Cross-Modal Attention Based
  Audio-Visual Conformer
Robust Wake Word Spotting With Frame-Level Cross-Modal Attention Based Audio-Visual Conformer
Haoxu Wang
Ming Cheng
Qiang Fu
Ming Li
47
1
0
04 Mar 2024
The Hidden Attention of Mamba Models
The Hidden Attention of Mamba Models
Ameen Ali
Itamar Zimerman
Lior Wolf
Mamba
39
58
0
03 Mar 2024
ProtoP-OD: Explainable Object Detection with Prototypical Parts
ProtoP-OD: Explainable Object Detection with Prototypical Parts
Pavlos Rath-Manakidis
Frederik Strothmann
Tobias Glasmachers
Laurenz Wiskott
ViT
35
1
0
29 Feb 2024
Evaluating Webcam-based Gaze Data as an Alternative for Human Rationale
  Annotations
Evaluating Webcam-based Gaze Data as an Alternative for Human Rationale Annotations
Stephanie Brandl
Oliver Eberle
Tiago F. R. Ribeiro
Anders Søgaard
Nora Hollenstein
40
1
0
29 Feb 2024
What Text Design Characterizes Book Genres?
What Text Design Characterizes Book Genres?
Daichi Haraguchi
Brian Kenji Iwana
Seiichi Uchida
19
2
0
26 Feb 2024
Attention-aware Semantic Communications for Collaborative Inference
Attention-aware Semantic Communications for Collaborative Inference
Jiwoong Im
Nayoung Kwon
Taewoo Park
Jiheon Woo
Jaeho Lee
Yongjune Kim
46
2
0
23 Feb 2024
When Only Time Will Tell: Interpreting How Transformers Process Local
  Ambiguities Through the Lens of Restart-Incrementality
When Only Time Will Tell: Interpreting How Transformers Process Local Ambiguities Through the Lens of Restart-Incrementality
Brielen Madureira
Patrick Kahardipraja
David Schlangen
39
2
0
20 Feb 2024
AttnLRP: Attention-Aware Layer-Wise Relevance Propagation for
  Transformers
AttnLRP: Attention-Aware Layer-Wise Relevance Propagation for Transformers
Reduan Achtibat
Sayed Mohammad Vakilzadeh Hatefi
Maximilian Dreyer
Aakriti Jain
Thomas Wiegand
Sebastian Lapuschkin
Wojciech Samek
33
25
0
08 Feb 2024
Leveraging Swin Transformer for Local-to-Global Weakly Supervised
  Semantic Segmentation
Leveraging Swin Transformer for Local-to-Global Weakly Supervised Semantic Segmentation
Rozhan Ahmadi
S. Kasaei
ViT
38
2
0
31 Jan 2024
Stochastic Amortization: A Unified Approach to Accelerate Feature and
  Data Attribution
Stochastic Amortization: A Unified Approach to Accelerate Feature and Data Attribution
Ian Covert
Chanwoo Kim
Su-In Lee
James Zou
Tatsunori Hashimoto
TDI
35
8
0
29 Jan 2024
From Understanding to Utilization: A Survey on Explainability for Large
  Language Models
From Understanding to Utilization: A Survey on Explainability for Large Language Models
Haoyan Luo
Lucia Specia
56
20
0
23 Jan 2024
Better Explain Transformers by Illuminating Important Information
Better Explain Transformers by Illuminating Important Information
Linxin Song
Yan Cui
Ao Luo
Freddy Lecue
Irene Z Li
FAtt
28
1
0
18 Jan 2024
B-Cos Aligned Transformers Learn Human-Interpretable Features
B-Cos Aligned Transformers Learn Human-Interpretable Features
Manuel Tran
Amal Lahiani
Yashin Dicente Cid
Melanie Boxberg
Peter Lienemann
C. Matek
S. J. Wagner
Fabian J. Theis
Eldad Klaiman
Tingying Peng
MedIm
ViT
21
2
0
16 Jan 2024
Statistical Test for Attention Map in Vision Transformer
Statistical Test for Attention Map in Vision Transformer
Tomohiro Shiraishi
Daiki Miwa
Teruyuki Katsuoka
Vo Nguyen Le Duy
Koichi Taji
Ichiro Takeuchi
32
5
0
16 Jan 2024
Towards Explainable Artificial Intelligence (XAI): A Data Mining
  Perspective
Towards Explainable Artificial Intelligence (XAI): A Data Mining Perspective
Haoyi Xiong
Xuhong Li
Xiaofei Zhang
Jiamin Chen
Xinhao Sun
Yuchen Li
Zeyi Sun
Jundong Li
XAI
40
8
0
09 Jan 2024
Identifying Important Group of Pixels using Interactions
Identifying Important Group of Pixels using Interactions
Kosuke Sumiyasu
Kazuhiko Kawamoto
Hiroshi Kera
21
2
0
08 Jan 2024
XAI for In-hospital Mortality Prediction via Multimodal ICU Data
XAI for In-hospital Mortality Prediction via Multimodal ICU Data
Xingqiao Li
Jindong Gu
Zhiyong Wang
Yancheng Yuan
Bo Du
Fengxiang He
32
2
0
29 Dec 2023
DRStageNet: Deep Learning for Diabetic Retinopathy Staging from Fundus
  Images
DRStageNet: Deep Learning for Diabetic Retinopathy Staging from Fundus Images
Yevgeniy Men
Jonathan Fhima
Leo Anthony Celi
L. Z. Ribeiro
Luis Filipe Nakayama
Joachim A. Behar
38
4
0
22 Dec 2023
Explainable Multi-Camera 3D Object Detection with Transformer-Based
  Saliency Maps
Explainable Multi-Camera 3D Object Detection with Transformer-Based Saliency Maps
Till Beemelmanns
Wassim Zahr
Lutz Eckstein
32
0
0
22 Dec 2023
Improved Visual Grounding through Self-Consistent Explanations
Improved Visual Grounding through Self-Consistent Explanations
Ruozhen He
Paola Cascante-Bonilla
Ziyan Yang
Alexander C. Berg
Vicente Ordonez
ReLM
ObjD
LRM
FAtt
35
8
0
07 Dec 2023
Class-Discriminative Attention Maps for Vision Transformers
Class-Discriminative Attention Maps for Vision Transformers
L. Brocki
Jakub Binda
N. C. Chung
MedIm
32
3
0
04 Dec 2023
Previous
123456789
Next