Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2107.05768
Cited By
Combiner: Full Attention Transformer with Sparse Computation Cost
12 July 2021
Hongyu Ren
H. Dai
Zihang Dai
Mengjiao Yang
J. Leskovec
Dale Schuurmans
Bo Dai
Re-assign community
ArXiv
PDF
HTML
Papers citing
"Combiner: Full Attention Transformer with Sparse Computation Cost"
5 / 5 papers shown
Title
Dual Filter: A Mathematical Framework for Inference using Transformer-like Architectures
Heng-Sheng Chang
P. Mehta
29
0
0
01 May 2025
Segmentation Guided Sparse Transformer for Under-Display Camera Image Restoration
Jingyun Xue
Tao Wang
Jun Wang
Kaihao Zhang
ViT
20
2
0
09 Mar 2024
Big Bird: Transformers for Longer Sequences
Manzil Zaheer
Guru Guruganesh
Kumar Avinava Dubey
Joshua Ainslie
Chris Alberti
...
Philip Pham
Anirudh Ravula
Qifan Wang
Li Yang
Amr Ahmed
VLM
243
1,684
0
28 Jul 2020
Efficient Content-Based Sparse Attention with Routing Transformers
Aurko Roy
M. Saffar
Ashish Vaswani
David Grangier
MoE
225
502
0
12 Mar 2020
Pixel Recurrent Neural Networks
Aaron van den Oord
Nal Kalchbrenner
Koray Kavukcuoglu
SSeg
GAN
216
2,391
0
25 Jan 2016
1