Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
1802.05751
Cited By
Image Transformer
15 February 2018
Niki Parmar
Ashish Vaswani
Jakob Uszkoreit
Lukasz Kaiser
Noam M. Shazeer
Alexander Ku
Dustin Tran
ViT
Re-assign community
ArXiv
PDF
HTML
Papers citing
"Image Transformer"
50 / 277 papers shown
Title
ASSET: Autoregressive Semantic Scene Editing with Transformers at High Resolutions
Difan Liu
Sandesh Shetty
Tobias Hinz
Matthew Fisher
Richard Y. Zhang
Taesung Park
E. Kalogerakis
ViT
27
29
0
24 May 2022
BodyMap: Learning Full-Body Dense Correspondence Map
A. Ianina
N. Sarafianos
Yuanlu Xu
Ignacio Rocco
Tony Tung
3DH
22
14
0
18 May 2022
MulT: An End-to-End Multitask Learning Transformer
Deblina Bhattacharjee
Tong Zhang
Sabine Süsstrunk
Mathieu Salzmann
ViT
34
62
0
17 May 2022
Attention Mechanism in Neural Networks: Where it Comes and Where it Goes
Derya Soydaner
3DV
36
149
0
27 Apr 2022
ClusterGNN: Cluster-based Coarse-to-Fine Graph Neural Network for Efficient Feature Matching
Yanxing Shi
Junxiong Cai
Yoli Shavit
Tai-Jiang Mu
Wensen Feng
Kai Zhang
GNN
19
77
0
25 Apr 2022
Efficient Linear Attention for Fast and Accurate Keypoint Matching
Suwichaya Suwanwimolkul
S. Komorita
3DPC
3DV
19
11
0
16 Apr 2022
A Call for Clarity in Beam Search: How It Works and When It Stops
Jungo Kasai
Keisuke Sakaguchi
Ronan Le Bras
Dragomir R. Radev
Yejin Choi
Noah A. Smith
26
6
0
11 Apr 2022
Towards An End-to-End Framework for Flow-Guided Video Inpainting
Z. Li
Cheng Lu
Jia Qin
Chunle Guo
Mingg-Ming Cheng
41
149
0
06 Apr 2022
FoV-Net: Field-of-View Extrapolation Using Self-Attention and Uncertainty
Liqian Ma
Stamatios Georgoulis
Xu Jia
Luc Van Gool
24
6
0
04 Apr 2022
Unitail: Detecting, Reading, and Matching in Retail Scene
Fangyi Chen
Han Zhang
Zaiwang Li
Jiachen Dou
Shentong Mo
Hao Chen
Yongxin Zhang
Uzair Ahmed
Chenchen Zhu
Marios Savvides
22
9
0
01 Apr 2022
Domain Invariant Siamese Attention Mask for Small Object Change Detection via Everyday Indoor Robot Navigation
Koji Takeda
Kanji Tanaka
Yoshimasa Nakamura
3DPC
11
3
0
29 Mar 2022
A General Survey on Attention Mechanisms in Deep Learning
Gianni Brauwers
Flavius Frasincar
28
296
0
27 Mar 2022
A World-Self Model Towards Understanding Intelligence
Yutao Yue
24
2
0
25 Mar 2022
Efficient-VDVAE: Less is more
Louay Hazami
Rayhane Mama
Ragavan Thurairatnam
BDL
19
28
0
25 Mar 2022
High-Performance Transformer Tracking
Xin Chen
B. Yan
Jiawen Zhu
Huchuan Lu
Xiang Ruan
D. Wang
ViT
21
33
0
25 Mar 2022
Make-A-Scene: Scene-Based Text-to-Image Generation with Human Priors
Oran Gafni
Adam Polyak
Oron Ashual
Shelly Sheynin
Devi Parikh
Yaniv Taigman
DiffM
17
510
0
24 Mar 2022
Linearizing Transformer with Key-Value Memory
Yizhe Zhang
Deng Cai
20
5
0
23 Mar 2022
ViewFormer: NeRF-free Neural Rendering from Few Images Using Transformers
Jonávs Kulhánek
Erik Derner
Torsten Sattler
Robert Babuvska
ViT
28
73
0
18 Mar 2022
Image Super-Resolution With Deep Variational Autoencoders
Darius Chira
Ilian Haralampiev
Ole Winther
Andrea Dittadi
Valentin Liévin
DRL
30
32
0
17 Mar 2022
Semantic-aligned Fusion Transformer for One-shot Object Detection
Yizhou Zhao
Xun Guo
Yan Lu
ViT
ObjD
20
21
0
17 Mar 2022
Deep Transformers Thirst for Comprehensive-Frequency Data
R. Xia
Chao Xue
Boyu Deng
Fang Wang
Jingchao Wang
ViT
25
0
0
14 Mar 2022
The Principle of Diversity: Training Stronger Vision Transformers Calls for Reducing All Levels of Redundancy
Tianlong Chen
Zhenyu (Allen) Zhang
Yu Cheng
Ahmed Hassan Awadallah
Zhangyang Wang
ViT
33
37
0
12 Mar 2022
Anti-Oversmoothing in Deep Vision Transformers via the Fourier Domain Analysis: From Theory to Practice
Peihao Wang
Wenqing Zheng
Tianlong Chen
Zhangyang Wang
ViT
22
127
0
09 Mar 2022
Signature and Log-signature for the Study of Empirical Distributions Generated with GANs
J. Curtò
I. D. Zarzà
Hong-Mei Yan
Carlos T. Calafate
16
1
0
07 Mar 2022
PanFormer: a Transformer Based Model for Pan-sharpening
Huanyu Zhou
Qingjie Liu
Yunhong Wang
ViT
20
42
0
06 Mar 2022
Contextformer: A Transformer with Spatio-Channel Attention for Context Modeling in Learned Image Compression
A. B. Koyuncu
Han Gao
Atanas Boev
Georgii Gaikov
Elena Alshina
Eckehard Steinbach
ViT
31
68
0
04 Mar 2022
Dynamic N:M Fine-grained Structured Sparse Attention Mechanism
Zhaodong Chen
Yuying Quan
Zheng Qu
L. Liu
Yufei Ding
Yuan Xie
28
22
0
28 Feb 2022
Ligandformer: A Graph Neural Network for Predicting Compound Property with Robust Interpretation
Jinjiang Guo
Qi Liu
Han Guo
Xi Lu
AI4CE
16
3
0
21 Feb 2022
TransDreamer: Reinforcement Learning with Transformer World Models
Changgu Chen
Yi-Fu Wu
Jaesik Yoon
Sungjin Ahn
OffRL
32
90
0
19 Feb 2022
MaskGIT: Masked Generative Image Transformer
Huiwen Chang
Han Zhang
Lu Jiang
Ce Liu
William T. Freeman
ViT
35
617
0
08 Feb 2022
LwPosr: Lightweight Efficient Fine-Grained Head Pose Estimation
Naina Dhingra
16
16
0
07 Feb 2022
Improving Sample Efficiency of Value Based Models Using Attention and Vision Transformers
Amir Ardalan Kalantari
Mohammad Amini
Sarath Chandar
Doina Precup
44
4
0
01 Feb 2022
Joint Liver and Hepatic Lesion Segmentation in MRI using a Hybrid CNN with Transformer Layers
Georg Hille
Shubham Agrawal
Pavan Tummala
C. Wybranski
M. Pech
A. Surov
S. Saalfeld
ViT
MedIm
11
26
0
26 Jan 2022
SA-VQA: Structured Alignment of Visual and Semantic Representations for Visual Question Answering
Peixi Xiong
Quanzeng You
Pei Yu
Zicheng Liu
Ying Wu
16
5
0
25 Jan 2022
RestoreFormer: High-Quality Blind Face Restoration from Undegraded Key-Value Pairs
Zhouxia Wang
Jiawei Zhang
Runjian Chen
Wenping Wang
Ping Luo
CVBM
13
109
0
17 Jan 2022
Continual Transformers: Redundancy-Free Attention for Online Inference
Lukas Hedegaard
Arian Bakhtiarnia
Alexandros Iosifidis
CLL
20
11
0
17 Jan 2022
DeepCreativity: Measuring Creativity with Deep Learning Techniques
Giorgio Franceschelli
Mirco Musolesi
26
5
0
16 Jan 2022
Video Transformers: A Survey
Javier Selva
A. S. Johansen
Sergio Escalera
Kamal Nasrollahi
T. Moeslund
Albert Clapés
ViT
22
103
0
16 Jan 2022
Learning Operators with Coupled Attention
Georgios Kissas
Jacob H. Seidman
Leonardo Ferreira Guilhoto
V. Preciado
George J. Pappas
P. Perdikaris
24
109
0
04 Jan 2022
Miti-DETR: Object Detection based on Transformers with Mitigatory Self-Attention Convergence
Wenchi Ma
Tianxiao Zhang
Guanghui Wang
ViT
28
14
0
26 Dec 2021
Efficient Visual Tracking with Exemplar Transformers
Philippe Blatter
Menelaos Kanakis
Martin Danelljan
Luc Van Gool
ViT
19
79
0
17 Dec 2021
EEG-Transformer: Self-attention from Transformer Architecture for Decoding EEG of Imagined Speech
Y. E. Lee
Seo-Hyun Lee
15
45
0
15 Dec 2021
Couplformer:Rethinking Vision Transformer with Coupling Attention Map
Hai Lan
Xihao Wang
Xian Wei
ViT
26
3
0
10 Dec 2021
FaceFormer: Speech-Driven 3D Facial Animation with Transformers
Yingruo Fan
Zhaojiang Lin
Jun Saito
Wenping Wang
Taku Komura
CVBM
38
194
0
10 Dec 2021
Trajectory-Constrained Deep Latent Visual Attention for Improved Local Planning in Presence of Heterogeneous Terrain
Stefan Wapnick
Travis Manderson
D. Meger
Gregory Dudek
31
5
0
09 Dec 2021
BEVT: BERT Pretraining of Video Transformers
Rui Wang
Dongdong Chen
Zuxuan Wu
Yinpeng Chen
Xiyang Dai
Mengchen Liu
Yu-Gang Jiang
Luowei Zhou
Lu Yuan
ViT
36
203
0
02 Dec 2021
EdiBERT, a generative model for image editing
Thibaut Issenhuth
Ugo Tanielian
Jérémie Mary
David Picard
DiffM
32
12
0
30 Nov 2021
Vector Quantized Diffusion Model for Text-to-Image Synthesis
Shuyang Gu
Dong Chen
Jianmin Bao
Fang Wen
Bo Zhang
Dongdong Chen
Lu Yuan
B. Guo
DiffM
45
757
0
29 Nov 2021
Point-BERT: Pre-training 3D Point Cloud Transformers with Masked Point Modeling
Xumin Yu
Lulu Tang
Yongming Rao
Tiejun Huang
Jie Zhou
Jiwen Lu
3DPC
27
652
0
29 Nov 2021
On the Integration of Self-Attention and Convolution
Xuran Pan
Chunjiang Ge
Rui Lu
S. Song
Guanfu Chen
Zeyi Huang
Gao Huang
SSL
36
287
0
29 Nov 2021
Previous
1
2
3
4
5
6
Next