Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2302.14705
Cited By
AccelTran: A Sparsity-Aware Accelerator for Dynamic Inference with Transformers
28 February 2023
Shikhar Tuli
N. Jha
Re-assign community
ArXiv
PDF
HTML
Papers citing
"AccelTran: A Sparsity-Aware Accelerator for Dynamic Inference with Transformers"
6 / 6 papers shown
Title
A Low-Power Streaming Speech Enhancement Accelerator For Edge Devices
Ci-Hao Wu
Tian-Sheuan Chang
61
1
0
27 Mar 2025
FedSpaLLM: Federated Pruning of Large Language Models
Guangji Bai
Yijiang Li
Zilinghan Li
Liang Zhao
Kibaek Kim
FedML
60
4
0
20 Feb 2025
A Heterogeneous Chiplet Architecture for Accelerating End-to-End Transformer Models
Harsh Sharma
Pratyush Dhingra
J. Doppa
Ümit Y. Ogras
P. Pande
32
7
0
18 Dec 2023
EdgeTran: Co-designing Transformers for Efficient Inference on Mobile Edge Platforms
Shikhar Tuli
N. Jha
34
3
0
24 Mar 2023
Energon: Towards Efficient Acceleration of Transformers Using Dynamic Sparse Attention
Zhe Zhou
Junling Liu
Zhenyu Gu
Guangyu Sun
56
42
0
18 Oct 2021
GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding
Alex Jinpeng Wang
Amanpreet Singh
Julian Michael
Felix Hill
Omer Levy
Samuel R. Bowman
ELM
294
6,943
0
20 Apr 2018
1