Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2110.11299
Cited By
Transformer Acceleration with Dynamic Sparse Attention
21 October 2021
Liu Liu
Zheng Qu
Zhaodong Chen
Yufei Ding
Yuan Xie
Re-assign community
ArXiv
PDF
HTML
Papers citing
"Transformer Acceleration with Dynamic Sparse Attention"
3 / 3 papers shown
Title
SeerAttention: Learning Intrinsic Sparse Attention in Your LLMs
Yizhao Gao
Zhichen Zeng
Dayou Du
Shijie Cao
Hayden Kwok-Hay So
...
Junjie Lai
Mao Yang
Ting Cao
Fan Yang
M. Yang
47
18
0
17 Oct 2024
Big Bird: Transformers for Longer Sequences
Manzil Zaheer
Guru Guruganesh
Kumar Avinava Dubey
Joshua Ainslie
Chris Alberti
...
Philip Pham
Anirudh Ravula
Qifan Wang
Li Yang
Amr Ahmed
VLM
249
2,009
0
28 Jul 2020
Efficient Content-Based Sparse Attention with Routing Transformers
Aurko Roy
M. Saffar
Ashish Vaswani
David Grangier
MoE
238
578
0
12 Mar 2020
1