Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
1904.09925
Cited By
Attention Augmented Convolutional Networks
22 April 2019
Irwan Bello
Barret Zoph
Ashish Vaswani
Jonathon Shlens
Quoc V. Le
Re-assign community
ArXiv
PDF
HTML
Papers citing
"Attention Augmented Convolutional Networks"
50 / 427 papers shown
Title
Local Multi-Head Channel Self-Attention for Facial Expression Recognition
Roberto Pecoraro
Valerio Basile
Viviana Bono
Sara Gallo
ViT
73
48
0
14 Nov 2021
Full-attention based Neural Architecture Search using Context Auto-regression
Yuan Zhou
Haiyang Wang
Shuwei Huo
Boyu Wang
25
3
0
13 Nov 2021
A Survey of Visual Transformers
Yang Liu
Yao Zhang
Yixin Wang
Feng Hou
Jin Yuan
Jiang Tian
Yang Zhang
Zhongchao Shi
Jianping Fan
Zhiqiang He
3DGS
ViT
69
330
0
11 Nov 2021
Are Transformers More Robust Than CNNs?
Yutong Bai
Jieru Mei
Alan Yuille
Cihang Xie
ViT
AAML
181
258
0
10 Nov 2021
Attention on Classification for Fire Segmentation
Milad Niknejad
Alexandre Bernardino
SSeg
17
6
0
04 Nov 2021
DPNET: Dual-Path Network for Efficient Object Detectioj with Lightweight Self-Attention
Huiming Shi
Quan Zhou
Yinghao Ni
Xiaofu Wu
Longin Jan Latecki
ObjD
14
8
0
31 Oct 2021
Dispensed Transformer Network for Unsupervised Domain Adaptation
Yunxiang Li
Jingxiong Li
Ruilong Dan
Shuai Wang
Kai Jin
...
Qianni Zhang
Huiyu Zhou
Qun Jin
Li Wang
Yaqi Wang
OOD
MedIm
15
4
0
28 Oct 2021
Geometric Transformer for End-to-End Molecule Properties Prediction
Yoni Choukroun
Lior Wolf
AI4CE
ViT
17
16
0
26 Oct 2021
IIP-Transformer: Intra-Inter-Part Transformer for Skeleton-Based Action Recognition
Qingtian Wang
Jianlin Peng
Shuze Shi
Tingxi Liu
Jiabin He
Renliang Weng
ViT
21
35
0
26 Oct 2021
Toward Accurate and Reliable Iris Segmentation Using Uncertainty Learning
Jianze Wei
Huaibo Huang
Muyi Sun
Yunlong Wang
Min Ren
R. He
Zhenan Sun
25
2
0
20 Oct 2021
MEDUSA: Multi-scale Encoder-Decoder Self-Attention Deep Neural Network Architecture for Medical Image Analysis
Hossein Aboutalebi
Maya Pavlova
Hayden Gunraj
M. Shafiee
A. Sabri
Amer Alaref
Alexander Wong
15
17
0
12 Oct 2021
Leveraging Transformers for StarCraft Macromanagement Prediction
Muhammad Junaid Khan
Shah Hassan
G. Sukthankar
17
5
0
11 Oct 2021
Context-LGM: Leveraging Object-Context Relation for Context-Aware Object Recognition
Mingzhou Liu
Xinwei Sun
Fandong Zhang
Yizhou Yu
Yizhou Wang
24
0
0
08 Oct 2021
MobileViT: Light-weight, General-purpose, and Mobile-friendly Vision Transformer
Sachin Mehta
Mohammad Rastegari
ViT
189
1,210
0
05 Oct 2021
Attention Augmented Convolutional Transformer for Tabular Time-series
Sharath M. Shankaranarayana
D. Runje
LMTD
AI4TS
58
8
0
05 Oct 2021
VTAMIQ: Transformers for Attention Modulated Image Quality Assessment
Andrei Chubarau
James Clark
ViT
30
9
0
04 Oct 2021
GT U-Net: A U-Net Like Group Transformer Network for Tooth Root Segmentation
Yunxiang Li
Shuai Wang
Jun Wang
G. Zeng
Wenjun Liu
Qianni Zhang
Qun Jin
Yaqi Wang
ViT
MedIm
23
47
0
30 Sep 2021
Improved Xception with Dual Attention Mechanism and Feature Fusion for Face Forgery Detection
Hao Lin
Weiqi Luo
Kangkang Wei
Minglin Liu
CVBM
AAML
3DPC
32
16
0
29 Sep 2021
Revisiting 3D ResNets for Video Recognition
Xianzhi Du
Yeqing Li
Yin Cui
Rui Qian
Jing Li
Irwan Bello
51
17
0
03 Sep 2021
Searching for Efficient Multi-Stage Vision Transformers
Yi-Lun Liao
S. Karaman
Vivienne Sze
ViT
16
19
0
01 Sep 2021
Efficient conformer: Progressive downsampling and grouped attention for automatic speech recognition
Maxime Burchi
Valentin Vielzeuf
23
84
0
31 Aug 2021
GroupFormer: Group Activity Recognition with Clustered Spatial-Temporal Transformer
Shuaicheng Li
Qianggang Cao
Lingbo Liu
Kunlin Yang
Shinan Liu
Jun Hou
Shuai Yi
ViT
34
102
0
28 Aug 2021
Learning Inner-Group Relations on Point Clouds
Haoxi Ran
Wei Zhuo
J. Liu
Li Lu
3DPC
31
59
0
27 Aug 2021
Self-Attention for Audio Super-Resolution
Nathanaël Carraz Rakotonirina
SupR
27
23
0
26 Aug 2021
Monocular Depth Estimation Primed by Salient Point Detection and Normalized Hessian Loss
Lam Huynh
Matteo Pedone
Phong H. Nguyen
Jirí Matas
Esa Rahtu
J. Heikkilä
MDE
3DPC
12
3
0
25 Aug 2021
Discovering Spatial Relationships by Transformers for Domain Generalization
Cuicui Kang
Karthik Nandakumar
ViT
11
0
0
23 Aug 2021
MM-ViT: Multi-Modal Video Transformer for Compressed Video Action Recognition
Jiawei Chen
C. Ho
ViT
24
76
0
20 Aug 2021
PoinTr: Diverse Point Cloud Completion with Geometry-Aware Transformers
Xumin Yu
Yongming Rao
Ziyi Wang
Zuyan Liu
Jiwen Lu
Jie Zhou
ViT
17
424
0
19 Aug 2021
Do Vision Transformers See Like Convolutional Neural Networks?
M. Raghu
Thomas Unterthiner
Simon Kornblith
Chiyuan Zhang
Alexey Dosovitskiy
ViT
20
922
0
19 Aug 2021
An Attention Module for Convolutional Neural Networks
Zhu Baozhou
P. Hofstee
Jinho Lee
Zaid Al-Ars
8
23
0
18 Aug 2021
No-Reference Image Quality Assessment via Transformers, Relative Ranking, and Self-Consistency
S. Golestaneh
Saba Dadsetan
Kris M. Kitani
ViT
6
241
0
16 Aug 2021
MUSIQ: Multi-scale Image Quality Transformer
Junjie Ke
Qifei Wang
Yilin Wang
P. Milanfar
Feng Yang
157
624
0
12 Aug 2021
TVT: Transferable Vision Transformer for Unsupervised Domain Adaptation
Jinyu Yang
Jingjing Liu
N. Xu
Junzhou Huang
17
125
0
12 Aug 2021
Learning Fair Face Representation With Progressive Cross Transformer
Yong Li
Yufei Sun
Zhen Cui
Shiguang Shan
Jian Yang
14
12
0
11 Aug 2021
RaftMLP: How Much Can Be Done Without Attention and with Less Spatial Locality?
Yuki Tatsunami
Masato Taki
19
12
0
09 Aug 2021
FA-GAN: Fused Attentive Generative Adversarial Networks for MRI Image Super-Resolution
M. Jiang
Min Zhi
Liying Wei
Xiaocheng Yang
Jucheng Zhang
Yongming Li
Pin Wang
Jiahao Huang
Guang Yang
MedIm
19
83
0
09 Aug 2021
Understanding the computational demands underlying visual reasoning
Mohit Vaishnav
Rémi Cadène
A. Alamia
Drew Linsley
Rufin VanRullen
Thomas Serre
GNN
CoGe
32
16
0
08 Aug 2021
Unifying Global-Local Representations in Salient Object Detection with Transformer
Sucheng Ren
Qiang Wen
Nanxuan Zhao
Guoqiang Han
Shengfeng He
ViT
26
25
0
05 Aug 2021
Vision Transformer with Progressive Sampling
Xiaoyu Yue
Shuyang Sun
Zhanghui Kuang
Meng Wei
Philip H. S. Torr
Wayne Zhang
Dahua Lin
ViT
16
81
0
03 Aug 2021
Contextual Transformer Networks for Visual Recognition
Yehao Li
Ting Yao
Yingwei Pan
Tao Mei
ViT
8
468
0
26 Jul 2021
H-Transformer-1D: Fast One-Dimensional Hierarchical Attention for Sequences
Zhenhai Zhu
Radu Soricut
98
41
0
25 Jul 2021
QVHighlights: Detecting Moments and Highlights in Videos via Natural Language Queries
Jie Lei
Tamara L. Berg
Mohit Bansal
ViT
19
62
0
20 Jul 2021
Attention-Guided NIR Image Colorization via Adaptive Fusion of Semantic and Texture Clues
Xingxing Yang
Jie Chen
Zaifeng Yang
Zhenghua Chen
DiffM
11
2
0
20 Jul 2021
All the attention you need: Global-local, spatial-channel attention for image retrieval
Chull Hwan Song
Hye Joo Han
Yannis Avrithis
11
39
0
16 Jul 2021
Visual Parser: Representing Part-whole Hierarchies with Transformers
Shuyang Sun
Xiaoyu Yue
S. Bai
Philip H. S. Torr
50
27
0
13 Jul 2021
Locally Enhanced Self-Attention: Combining Self-Attention and Convolution as Local and Context Terms
Chenglin Yang
Siyuan Qiao
Adam Kortylewski
Alan Yuille
14
4
0
12 Jul 2021
Local-to-Global Self-Attention in Vision Transformers
Jinpeng Li
Yichao Yan
Shengcai Liao
Xiaokang Yang
Ling Shao
ViT
17
29
0
10 Jul 2021
Poly-NL: Linear Complexity Non-local Layers with Polynomials
F. Babiloni
Ioannis Marras
Filippos Kokkinos
Jiankang Deng
Grigorios G. Chrysos
S. Zafeiriou
31
6
0
06 Jul 2021
Polarized Self-Attention: Towards High-quality Pixel-wise Regression
Huajun Liu
Fuqiang Liu
Xinyi Fan
Dong Huang
72
210
0
02 Jul 2021
UTNet: A Hybrid Transformer Architecture for Medical Image Segmentation
Yunhe Gao
Mu Zhou
Dimitris N. Metaxas
MedIm
ViT
11
425
0
02 Jul 2021
Previous
1
2
3
4
5
6
7
8
9
Next