Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2305.19370
Cited By
Blockwise Parallel Transformer for Large Context Models
30 May 2023
Hao Liu
Pieter Abbeel
Re-assign community
ArXiv
PDF
HTML
Papers citing
"Blockwise Parallel Transformer for Large Context Models"
4 / 4 papers shown
Title
Resurrecting Recurrent Neural Networks for Long Sequences
Antonio Orvieto
Samuel L. Smith
Albert Gu
Anushan Fernando
Çağlar Gülçehre
Razvan Pascanu
Soham De
88
258
0
11 Mar 2023
Transformer Quality in Linear Time
Weizhe Hua
Zihang Dai
Hanxiao Liu
Quoc V. Le
71
220
0
21 Feb 2022
LambdaNetworks: Modeling Long-Range Interactions Without Attention
Irwan Bello
260
178
0
17 Feb 2021
Megatron-LM: Training Multi-Billion Parameter Language Models Using Model Parallelism
M. Shoeybi
M. Patwary
Raul Puri
P. LeGresley
Jared Casper
Bryan Catanzaro
MoE
243
1,791
0
17 Sep 2019
1