ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2004.13621
  4. Cited By
Exploring Self-attention for Image Recognition

Exploring Self-attention for Image Recognition

28 April 2020
Hengshuang Zhao
Jiaya Jia
V. Koltun
    SSL
ArXivPDFHTML

Papers citing "Exploring Self-attention for Image Recognition"

50 / 316 papers shown
Title
Finding Strong Gravitational Lenses Through Self-Attention
Finding Strong Gravitational Lenses Through Self-Attention
H. Thuruthipilly
A. Zadrożny
Agnieszka Pollo
Marek Biesiada
16
6
0
18 Oct 2021
Attention meets Geometry: Geometry Guided Spatial-Temporal Attention for
  Consistent Self-Supervised Monocular Depth Estimation
Attention meets Geometry: Geometry Guided Spatial-Temporal Attention for Consistent Self-Supervised Monocular Depth Estimation
Patrick Ruhkamp
Daoyi Gao
Hanzhi Chen
Nassir Navab
Benjamin Busam
ViT
MDE
77
36
0
15 Oct 2021
Development and testing of an image transformer for explainable
  autonomous driving systems
Development and testing of an image transformer for explainable autonomous driving systems
Jiqian Dong
Sikai Chen
Shuya Zong
Tiantian Chen
Mohammad Miralinaghi
S. Labi
ViT
18
23
0
11 Oct 2021
Attentive Walk-Aggregating Graph Neural Networks
Attentive Walk-Aggregating Graph Neural Networks
M. F. Demirel
Shengchao Liu
Siddhant Garg
Zhenmei Shi
Yingyu Liang
87
9
0
06 Oct 2021
LEA-Net: Layer-wise External Attention Network for Efficient Color
  Anomaly Detection
LEA-Net: Layer-wise External Attention Network for Efficient Color Anomaly Detection
Ryoya Katafuchi
T. Tokunaga
14
3
0
12 Sep 2021
Pose-guided Inter- and Intra-part Relational Transformer for Occluded
  Person Re-Identification
Pose-guided Inter- and Intra-part Relational Transformer for Occluded Person Re-Identification
Zhongxing Ma
Yifan Zhao
Jia Li
ViT
16
53
0
08 Sep 2021
Impact of Attention on Adversarial Robustness of Image Classification
  Models
Impact of Attention on Adversarial Robustness of Image Classification Models
Prachi Agrawal
Narinder Singh Punn
S. K. Sonbhadra
Sonali Agarwal
AAML
16
6
0
02 Sep 2021
Searching for Efficient Multi-Stage Vision Transformers
Searching for Efficient Multi-Stage Vision Transformers
Yi-Lun Liao
S. Karaman
Vivienne Sze
ViT
16
19
0
01 Sep 2021
Learning Inner-Group Relations on Point Clouds
Learning Inner-Group Relations on Point Clouds
Haoxi Ran
Wei Zhuo
J. Liu
Li Lu
3DPC
33
59
0
27 Aug 2021
Relational Embedding for Few-Shot Classification
Relational Embedding for Few-Shot Classification
Dahyun Kang
Heeseung Kwon
Juhong Min
Minsu Cho
26
185
0
22 Aug 2021
PTT: Point-Track-Transformer Module for 3D Single Object Tracking in
  Point Clouds
PTT: Point-Track-Transformer Module for 3D Single Object Tracking in Point Clouds
Jiayao Shan
Sifan Zhou
Zheng Fang
Yubo Cui
ViT
17
79
0
14 Aug 2021
PVT: Point-Voxel Transformer for Point Cloud Learning
PVT: Point-Voxel Transformer for Point Cloud Learning
Cheng Zhang
Haocheng Wan
Xinyi Shen
Zizhao Wu
3DPC
ViT
11
80
0
13 Aug 2021
Transductive Few-Shot Classification on the Oblique Manifold
Transductive Few-Shot Classification on the Oblique Manifold
Guodong Qi
Huimin Yu
Zhaohui Lu
Shuzhao Li
11
45
0
09 Aug 2021
Unifying Global-Local Representations in Salient Object Detection with
  Transformer
Unifying Global-Local Representations in Salient Object Detection with Transformer
Sucheng Ren
Qiang Wen
Nanxuan Zhao
Guoqiang Han
Shengfeng He
ViT
26
25
0
05 Aug 2021
Contextual Transformer Networks for Visual Recognition
Contextual Transformer Networks for Visual Recognition
Yehao Li
Ting Yao
Yingwei Pan
Tao Mei
ViT
16
468
0
26 Jul 2021
All the attention you need: Global-local, spatial-channel attention for
  image retrieval
All the attention you need: Global-local, spatial-channel attention for image retrieval
Chull Hwan Song
Hye Joo Han
Yannis Avrithis
11
39
0
16 Jul 2021
Visual Parser: Representing Part-whole Hierarchies with Transformers
Visual Parser: Representing Part-whole Hierarchies with Transformers
Shuyang Sun
Xiaoyu Yue
S. Bai
Philip H. S. Torr
50
27
0
13 Jul 2021
Locally Enhanced Self-Attention: Combining Self-Attention and
  Convolution as Local and Context Terms
Locally Enhanced Self-Attention: Combining Self-Attention and Convolution as Local and Context Terms
Chenglin Yang
Siyuan Qiao
Adam Kortylewski
Alan Yuille
20
4
0
12 Jul 2021
GLiT: Neural Architecture Search for Global and Local Image Transformer
GLiT: Neural Architecture Search for Global and Local Image Transformer
Boyu Chen
Peixia Li
Chuming Li
Baopu Li
Lei Bai
Chen Lin
Ming-hui Sun
Junjie Yan
Wanli Ouyang
ViT
24
85
0
07 Jul 2021
Test-Time Personalization with a Transformer for Human Pose Estimation
Test-Time Personalization with a Transformer for Human Pose Estimation
Yizhuo Li
Miao Hao
Zonglin Di
N. B. Gundavarapu
Xiaolong Wang
ViT
25
45
0
05 Jul 2021
Learning Geometric Combinatorial Optimization Problems using
  Self-attention and Domain Knowledge
Learning Geometric Combinatorial Optimization Problems using Self-attention and Domain Knowledge
Jaeseung Lee
Woojin Choi
Jibum Kim
3DPC
12
0
0
05 Jul 2021
AutoFormer: Searching Transformers for Visual Recognition
AutoFormer: Searching Transformers for Visual Recognition
Minghao Chen
Houwen Peng
Jianlong Fu
Haibin Ling
ViT
36
259
0
01 Jul 2021
Early Convolutions Help Transformers See Better
Early Convolutions Help Transformers See Better
Tete Xiao
Mannat Singh
Eric Mintun
Trevor Darrell
Piotr Dollár
Ross B. Girshick
17
752
0
28 Jun 2021
VOLO: Vision Outlooker for Visual Recognition
VOLO: Vision Outlooker for Visual Recognition
Li-xin Yuan
Qibin Hou
Zihang Jiang
Jiashi Feng
Shuicheng Yan
ViT
46
313
0
24 Jun 2021
Probabilistic Attention for Interactive Segmentation
Probabilistic Attention for Interactive Segmentation
Prasad Gabbur
Manjot Bilkhu
J. Movellan
24
13
0
23 Jun 2021
TokenLearner: What Can 8 Learned Tokens Do for Images and Videos?
TokenLearner: What Can 8 Learned Tokens Do for Images and Videos?
Michael S. Ryoo
A. Piergiovanni
Anurag Arnab
Mostafa Dehghani
A. Angelova
ViT
23
127
0
21 Jun 2021
Visual Correspondence Hallucination
Visual Correspondence Hallucination
Hugo Germain
Vincent Lepetit
Guillaume Bourmaud
28
11
0
17 Jun 2021
XCiT: Cross-Covariance Image Transformers
XCiT: Cross-Covariance Image Transformers
Alaaeldin El-Nouby
Hugo Touvron
Mathilde Caron
Piotr Bojanowski
Matthijs Douze
...
Ivan Laptev
Natalia Neverova
Gabriel Synnaeve
Jakob Verbeek
Hervé Jégou
ViT
23
497
0
17 Jun 2021
Transformed CNNs: recasting pre-trained convolutional layers with
  self-attention
Transformed CNNs: recasting pre-trained convolutional layers with self-attention
Stéphane dÁscoli
Levent Sagun
Giulio Biroli
Ari S. Morcos
ViT
8
6
0
10 Jun 2021
On the Connection between Local Attention and Dynamic Depth-wise
  Convolution
On the Connection between Local Attention and Dynamic Depth-wise Convolution
Qi Han
Zejia Fan
Qi Dai
Lei-huan Sun
Ming-Ming Cheng
Jiaying Liu
Jingdong Wang
ViT
16
105
0
08 Jun 2021
Signal Transformer: Complex-valued Attention and Meta-Learning for
  Signal Recognition
Signal Transformer: Complex-valued Attention and Meta-Learning for Signal Recognition
Yihong Dong
Ying Peng
Muqiao Yang
Songtao Lu
Qingjiang Shi
38
9
0
05 Jun 2021
RegionViT: Regional-to-Local Attention for Vision Transformers
RegionViT: Regional-to-Local Attention for Vision Transformers
Chun-Fu Chen
Rameswar Panda
Quanfu Fan
ViT
16
193
0
04 Jun 2021
Glance-and-Gaze Vision Transformer
Glance-and-Gaze Vision Transformer
Qihang Yu
Yingda Xia
Yutong Bai
Yongyi Lu
Alan Yuille
Wei Shen
ViT
10
74
0
04 Jun 2021
Multi-Scale Feature Aggregation by Cross-Scale Pixel-to-Region Relation
  Operation for Semantic Segmentation
Multi-Scale Feature Aggregation by Cross-Scale Pixel-to-Region Relation Operation for Semantic Segmentation
Yechao Bai
Ziyuan Huang
Lyuyu Shen
Hongliang Guo
Marcelo H. Ang Jr
Daniela Rus
SSeg
11
4
0
03 Jun 2021
DLA-Net: Learning Dual Local Attention Features for Semantic
  Segmentation of Large-Scale Building Facade Point Clouds
DLA-Net: Learning Dual Local Attention Features for Semantic Segmentation of Large-Scale Building Facade Point Clouds
Yanfei Su
Weiquan Liu
Zhimin Yuan
Ming Cheng
Zhihong Zhang
Xuelun Shen
Cheng-Yu Wang
3DPC
12
38
0
01 Jun 2021
MSG-Transformer: Exchanging Local Spatial Information by Manipulating
  Messenger Tokens
MSG-Transformer: Exchanging Local Spatial Information by Manipulating Messenger Tokens
Jiemin Fang
Lingxi Xie
Xinggang Wang
Xiaopeng Zhang
Wenyu Liu
Qi Tian
ViT
13
73
0
31 May 2021
KVT: k-NN Attention for Boosting Vision Transformers
KVT: k-NN Attention for Boosting Vision Transformers
Pichao Wang
Xue Wang
F. Wang
Ming Lin
Shuning Chang
Hao Li
R. L. Jin
ViT
32
105
0
28 May 2021
Interpretable UAV Collision Avoidance using Deep Reinforcement Learning
Interpretable UAV Collision Avoidance using Deep Reinforcement Learning
Deep Thomas
Daniil Olshanskyi
Karter Krueger
Tichakorn Wongpiromsarn
Ali Jannesari
11
5
0
25 May 2021
Content-Augmented Feature Pyramid Network with Light Linear Spatial
  Transformers for Object Detection
Content-Augmented Feature Pyramid Network with Light Linear Spatial Transformers for Object Detection
Yongxiang Gu
Xiaolin Qin
Yuncong Peng
Lu Li
ViT
8
6
0
20 May 2021
SCTN: Sparse Convolution-Transformer Network for Scene Flow Estimation
SCTN: Sparse Convolution-Transformer Network for Scene Flow Estimation
Bing Li
Cheng Zheng
Silvio Giancola
Bernard Ghanem
ViT
3DPC
13
44
0
10 May 2021
Graph Attention Networks with Positional Embeddings
Graph Attention Networks with Positional Embeddings
Liheng Ma
Reihaneh Rabbany
Adriana Romero Soriano
GNN
17
20
0
09 May 2021
ResMLP: Feedforward networks for image classification with
  data-efficient training
ResMLP: Feedforward networks for image classification with data-efficient training
Hugo Touvron
Piotr Bojanowski
Mathilde Caron
Matthieu Cord
Alaaeldin El-Nouby
...
Gautier Izacard
Armand Joulin
Gabriel Synnaeve
Jakob Verbeek
Hervé Jégou
VLM
16
655
0
07 May 2021
Beyond Self-attention: External Attention using Two Linear Layers for
  Visual Tasks
Beyond Self-attention: External Attention using Two Linear Layers for Visual Tasks
Meng-Hao Guo
Zheng-Ning Liu
Tai-Jiang Mu
Shimin Hu
20
473
0
05 May 2021
Dual-Cross Central Difference Network for Face Anti-Spoofing
Dual-Cross Central Difference Network for Face Anti-Spoofing
Zitong Yu
Yunxiao Qin
Hengshuang Zhao
Xiaobai Li
Guoying Zhao
CVBM
17
103
0
04 May 2021
Emerging Properties in Self-Supervised Vision Transformers
Emerging Properties in Self-Supervised Vision Transformers
Mathilde Caron
Hugo Touvron
Ishan Misra
Hervé Jégou
Julien Mairal
Piotr Bojanowski
Armand Joulin
308
5,773
0
29 Apr 2021
ConTNet: Why not use convolution and transformer at the same time?
ConTNet: Why not use convolution and transformer at the same time?
Haotian Yan
Zhe Li
Weijian Li
Changhu Wang
Ming Wu
Chuang Zhang
ViT
12
76
0
27 Apr 2021
If your data distribution shifts, use self-learning
If your data distribution shifts, use self-learning
E. Rusak
Steffen Schneider
George Pachitariu
L. Eck
Peter V. Gehler
Oliver Bringmann
Wieland Brendel
Matthias Bethge
VLM
OOD
TTA
77
29
0
27 Apr 2021
Visformer: The Vision-friendly Transformer
Visformer: The Vision-friendly Transformer
Zhengsu Chen
Lingxi Xie
Jianwei Niu
Xuefeng Liu
Longhui Wei
Qi Tian
ViT
111
209
0
26 Apr 2021
AttWalk: Attentive Cross-Walks for Deep Mesh Analysis
AttWalk: Attentive Cross-Walks for Deep Mesh Analysis
Ran Ben Izhak
Alon Lahav
A. Tal
3DV
29
10
0
23 Apr 2021
Multiscale Vision Transformers
Multiscale Vision Transformers
Haoqi Fan
Bo Xiong
K. Mangalam
Yanghao Li
Zhicheng Yan
Jitendra Malik
Christoph Feichtenhofer
ViT
21
1,221
0
22 Apr 2021
Previous
1234567
Next