Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2005.14187
Cited By
HAT: Hardware-Aware Transformers for Efficient Natural Language Processing
28 May 2020
Hanrui Wang
Zhanghao Wu
Zhijian Liu
Han Cai
Ligeng Zhu
Chuang Gan
Song Han
Re-assign community
ArXiv
PDF
HTML
Papers citing
"HAT: Hardware-Aware Transformers for Efficient Natural Language Processing"
6 / 56 papers shown
Title
SpAtten: Efficient Sparse Attention Architecture with Cascade Token and Head Pruning
Hanrui Wang
Zhekai Zhang
Song Han
20
373
0
17 Dec 2020
Efficient Transformers: A Survey
Yi Tay
Mostafa Dehghani
Dara Bahri
Donald Metzler
VLM
74
1,101
0
14 Sep 2020
Searching Efficient 3D Architectures with Sparse Point-Voxel Convolution
Haotian Tang
Zhijian Liu
Shengyu Zhao
Yujun Lin
Ji Lin
Hanrui Wang
Song Han
3DPC
28
630
0
31 Jul 2020
GOBO: Quantizing Attention-Based NLP Models for Low Latency and Energy Efficient Inference
Ali Hadi Zadeh
Isak Edo
Omar Mohamed Awad
Andreas Moshovos
MQ
22
183
0
08 May 2020
Neural Architecture Search with Reinforcement Learning
Barret Zoph
Quoc V. Le
264
5,326
0
05 Nov 2016
Convolutional Neural Networks for Sentence Classification
Yoon Kim
AILaw
VLM
255
13,364
0
25 Aug 2014
Previous
1
2