ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2104.05704
  4. Cited By
Escaping the Big Data Paradigm with Compact Transformers

Escaping the Big Data Paradigm with Compact Transformers

12 April 2021
Ali Hassani
Steven Walton
Nikhil Shah
Abulikemu Abuduweili
Jiachen Li
Humphrey Shi
ArXivPDFHTML

Papers citing "Escaping the Big Data Paradigm with Compact Transformers"

15 / 215 papers shown
Title
Escaping the Gradient Vanishing: Periodic Alternatives of Softmax in
  Attention Mechanism
Escaping the Gradient Vanishing: Periodic Alternatives of Softmax in Attention Mechanism
Shulun Wang
Bin Liu
Feng Liu
9
16
0
16 Aug 2021
TriTransNet: RGB-D Salient Object Detection with a Triplet Transformer
  Embedding Network
TriTransNet: RGB-D Salient Object Detection with a Triplet Transformer Embedding Network
Zhengyi Liu
Yuan Wang
Zhengzheng Tu
Yun Xiao
Bin Tang
ViT
16
142
0
09 Aug 2021
Vision Transformer for femur fracture classification
Vision Transformer for femur fracture classification
L. Tanzi
A. Audisio
G. Cirrincione
A. Aprato
E. Vezzetti
MedIm
20
64
0
07 Aug 2021
Vision Xformers: Efficient Attention for Image Classification
Vision Xformers: Efficient Attention for Image Classification
Pranav Jeevan
Amit Sethi
ViT
14
13
0
05 Jul 2021
MSN: Efficient Online Mask Selection Network for Video Instance
  Segmentation
MSN: Efficient Online Mask Selection Network for Video Instance Segmentation
Vidit Goel
Jiachen Li
Shubhika Garg
Harsh Maheshwari
Humphrey Shi
17
7
0
19 Jun 2021
On Deep Neural Network Calibration by Regularization and its Impact on Refinement
Aditya Singh
Alessandro Bay
B. Sengupta
Andrea Mirabile
AAML
17
2
0
17 Jun 2021
Shuffle Transformer: Rethinking Spatial Shuffle for Vision Transformer
Shuffle Transformer: Rethinking Spatial Shuffle for Vision Transformer
Zilong Huang
Youcheng Ben
Guozhong Luo
Pei Cheng
Gang Yu
Bin-Bin Fu
ViT
11
181
0
07 Jun 2021
A Little Robustness Goes a Long Way: Leveraging Robust Features for
  Targeted Transfer Attacks
A Little Robustness Goes a Long Way: Leveraging Robust Features for Targeted Transfer Attacks
Jacob Mitchell Springer
Melanie Mitchell
Garrett T. Kenyon
AAML
12
43
0
03 Jun 2021
Nested Hierarchical Transformer: Towards Accurate, Data-Efficient and
  Interpretable Visual Understanding
Nested Hierarchical Transformer: Towards Accurate, Data-Efficient and Interpretable Visual Understanding
Zizhao Zhang
Han Zhang
Long Zhao
Ting Chen
Sercan Ö. Arik
Tomas Pfister
ViT
14
168
0
26 May 2021
Is Space-Time Attention All You Need for Video Understanding?
Is Space-Time Attention All You Need for Video Understanding?
Gedas Bertasius
Heng Wang
Lorenzo Torresani
ViT
278
1,978
0
09 Feb 2021
CheXtransfer: Performance and Parameter Efficiency of ImageNet Models
  for Chest X-Ray Interpretation
CheXtransfer: Performance and Parameter Efficiency of ImageNet Models for Chest X-Ray Interpretation
Alexander Ke
William Ellsworth
Oishi Banerjee
A. Ng
Pranav Rajpurkar
MedIm
63
101
0
18 Jan 2021
Transformers in Vision: A Survey
Transformers in Vision: A Survey
Salman Khan
Muzammal Naseer
Munawar Hayat
Syed Waqas Zamir
F. Khan
M. Shah
ViT
225
2,427
0
04 Jan 2021
Union-net: A deep neural network model adapted to small data sets
Union-net: A deep neural network model adapted to small data sets
Qingfang He
Guang Cheng
Zhiying Lin
PINN
26
6
0
24 Dec 2020
Stochastic-Sign SGD for Federated Learning with Theoretical Guarantees
Stochastic-Sign SGD for Federated Learning with Theoretical Guarantees
Richeng Jin
Yufan Huang
Xiaofan He
H. Dai
Tianfu Wu
FedML
4
63
0
25 Feb 2020
Effective Approaches to Attention-based Neural Machine Translation
Effective Approaches to Attention-based Neural Machine Translation
Thang Luong
Hieu H. Pham
Christopher D. Manning
214
7,687
0
17 Aug 2015
Previous
12345