Communities
Connect sessions
AI calendar
Organizations
Join Slack
Contact Sales
Search
Open menu
Home
Papers
All Papers
0 / 0 papers shown
Title
Home
Papers
2005.14187
Cited By
HAT: Hardware-Aware Transformers for Efficient Natural Language Processing
Annual Meeting of the Association for Computational Linguistics (ACL), 2020
28 May 2020
Hanrui Wang
Zhanghao Wu
Zhijian Liu
Han Cai
Ligeng Zhu
Chuang Gan
Song Han
Re-assign community
ArXiv (abs)
PDF
HTML
HuggingFace (2 upvotes)
Github (334★)
Papers citing
"HAT: Hardware-Aware Transformers for Efficient Natural Language Processing"
15 / 115 papers shown
Title
AlphaNet: Improved Training of Supernets with Alpha-Divergence
International Conference on Machine Learning (ICML), 2021
Dilin Wang
Chengyue Gong
Meng Li
Qiang Liu
Vikas Chandra
333
49
0
16 Feb 2021
Dancing along Battery: Enabling Transformer with Run-time Reconfigurability on Mobile Devices
Design Automation Conference (DAC), 2021
Yuhong Song
Weiwen Jiang
Bingbing Li
Panjie Qi
Qingfeng Zhuge
E. Sha
Sakyasingha Dasgupta
Yiyu Shi
Caiwen Ding
116
19
0
12 Feb 2021
A Comprehensive Survey on Hardware-Aware Neural Architecture Search
Hadjer Benmeziane
Kaoutar El Maghraoui
Hamza Ouarnoughi
Smail Niar
Martin Wistuba
Naigang Wang
170
121
0
22 Jan 2021
Transformers in Vision: A Survey
ACM Computing Surveys (CSUR), 2021
Salman Khan
Muzammal Naseer
Munawar Hayat
Syed Waqas Zamir
Fahad Shahbaz Khan
M. Shah
ViT
697
2,961
0
04 Jan 2021
SpAtten: Efficient Sparse Attention Architecture with Cascade Token and Head Pruning
International Symposium on High-Performance Computer Architecture (HPCA), 2020
Hanrui Wang
Zhekai Zhang
Song Han
291
466
0
17 Dec 2020
AttentiveNAS: Improving Neural Architecture Search via Attentive Sampling
Computer Vision and Pattern Recognition (CVPR), 2020
Dilin Wang
Meng Li
Chengyue Gong
Vikas Chandra
216
112
0
18 Nov 2020
Length-Adaptive Transformer: Train Once with Length Drop, Use Anytime with Search
Gyuwan Kim
Dong Wang
222
105
0
14 Oct 2020
Weight Distillation: Transferring the Knowledge in Neural Network Parameters
Annual Meeting of the Association for Computational Linguistics (ACL), 2020
Ye Lin
Yanyang Li
Ziyang Wang
Bei Li
Quan Du
Tong Xiao
Jingbo Zhu
230
28
0
19 Sep 2020
Efficient Transformers: A Survey
ACM Computing Surveys (ACM CSUR), 2020
Yi Tay
Mostafa Dehghani
Dara Bahri
Donald Metzler
VLM
580
1,279
0
14 Sep 2020
Pay Attention when Required
Swetha Mandava
Szymon Migacz
A. Fit-Florea
102
11
0
09 Sep 2020
Scaling Up Deep Neural Network Optimization for Edge Inference
IFIP International Information Security Conference (IIS), 2020
Bingqian Lu
Jianyi Yang
Shaolei Ren
244
4
0
01 Sep 2020
Searching Efficient 3D Architectures with Sparse Point-Voxel Convolution
European Conference on Computer Vision (ECCV), 2020
Haotian Tang
Zhijian Liu
Shengyu Zhao
Chengyue Wu
Ji Lin
Hanrui Wang
Song Han
3DPC
359
736
0
31 Jul 2020
MicroNet for Efficient Language Modeling
Zhongxia Yan
Hanrui Wang
Demi Guo
Song Han
159
8
0
16 May 2020
GOBO: Quantizing Attention-Based NLP Models for Low Latency and Energy Efficient Inference
Ali Hadi Zadeh
Isak Edo
Omar Mohamed Awad
Andreas Moshovos
MQ
194
205
0
08 May 2020
DynaBERT: Dynamic BERT with Adaptive Width and Depth
Neural Information Processing Systems (NeurIPS), 2020
Lu Hou
Zhiqi Huang
Lifeng Shang
Xin Jiang
Xiao Chen
Qun Liu
MQ
186
347
0
08 Apr 2020
Previous
1
2
3