ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2106.04803
  4. Cited By
CoAtNet: Marrying Convolution and Attention for All Data Sizes
v1v2 (latest)

CoAtNet: Marrying Convolution and Attention for All Data Sizes

Neural Information Processing Systems (NeurIPS), 2021
9 June 2021
Zihang Dai
Hanxiao Liu
Quoc V. Le
Mingxing Tan
    ViT
ArXiv (abs)PDFHTML

Papers citing "CoAtNet: Marrying Convolution and Attention for All Data Sizes"

10 / 510 papers shown
COVID-19 Pneumonia Severity Prediction using Hybrid
  Convolution-Attention Neural Architectures
COVID-19 Pneumonia Severity Prediction using Hybrid Convolution-Attention Neural Architectures
Nam H. Nguyen
Jerome Chang
183
3
0
06 Jul 2021
What Makes for Hierarchical Vision Transformer?
What Makes for Hierarchical Vision Transformer?
Yuxin Fang
Xinggang Wang
Rui Wu
Wenyu Liu
ViT
127
11
0
05 Jul 2021
Encoder-Decoder Architectures for Clinically Relevant Coronary Artery
  Segmentation
Encoder-Decoder Architectures for Clinically Relevant Coronary Artery SegmentationInternational Conference on Computational Advances in Bio and Medical Sciences (ICCABS), 2021
Joao Lourencco Silva
M. Menezes
T. Rodrigues
B. Silva
F. Pinto
Arlindo L. Oliveira
MedIm
187
22
0
21 Jun 2021
How to train your ViT? Data, Augmentation, and Regularization in Vision
  Transformers
How to train your ViT? Data, Augmentation, and Regularization in Vision Transformers
Andreas Steiner
Alexander Kolesnikov
Xiaohua Zhai
Ross Wightman
Jakob Uszkoreit
Lucas Beyer
ViT
335
765
0
18 Jun 2021
Scaling Vision Transformers
Scaling Vision TransformersComputer Vision and Pattern Recognition (CVPR), 2021
Xiaohua Zhai
Alexander Kolesnikov
N. Houlsby
Lucas Beyer
ViT
467
1,306
0
08 Jun 2021
Analogous to Evolutionary Algorithm: Designing a Unified Sequence Model
Analogous to Evolutionary Algorithm: Designing a Unified Sequence ModelNeural Information Processing Systems (NeurIPS), 2021
Jiangning Zhang
Chao Xu
Jian Li
Wenzhou Chen
Yabiao Wang
Ying Tai
Shuo Chen
Chengjie Wang
Feiyue Huang
Yong Liu
273
25
0
31 May 2021
Adversarial Robustness against Multiple and Single $l_p$-Threat Models
  via Quick Fine-Tuning of Robust Classifiers
Adversarial Robustness against Multiple and Single lpl_plp​-Threat Models via Quick Fine-Tuning of Robust ClassifiersInternational Conference on Machine Learning (ICML), 2021
Francesco Croce
Matthias Hein
OODAAML
210
26
0
26 May 2021
Visformer: The Vision-friendly Transformer
Visformer: The Vision-friendly TransformerIEEE International Conference on Computer Vision (ICCV), 2021
Zhengsu Chen
Lingxi Xie
Jianwei Niu
Xuefeng Liu
Longhui Wei
Qi Tian
ViT
499
271
0
26 Apr 2021
Conditional Positional Encodings for Vision Transformers
Conditional Positional Encodings for Vision TransformersInternational Conference on Learning Representations (ICLR), 2021
Xiangxiang Chu
Zhi Tian
Bo Zhang
Xinlong Wang
Chunhua Shen
ViT
399
740
0
22 Feb 2021
Transformers in Vision: A Survey
Transformers in Vision: A SurveyACM Computing Surveys (CSUR), 2021
Salman Khan
Muzammal Naseer
Munawar Hayat
Syed Waqas Zamir
Fahad Shahbaz Khan
M. Shah
ViT
914
3,125
0
04 Jan 2021
Previous
123...10119