Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2402.03009
Cited By
UniMem: Towards a Unified View of Long-Context Large Language Models
5 February 2024
Junjie Fang
Likai Tang
Hongzhe Bi
Yujia Qin
Si Sun
Zhenyu Li
Haolun Li
Yongjian Li
Xin Cong
Yukun Yan
Xiaodong Shi
Sen Song
Yankai Lin
Zhiyuan Liu
Maosong Sun
Re-assign community
ArXiv
PDF
HTML
Papers citing
"UniMem: Towards a Unified View of Long-Context Large Language Models"
4 / 4 papers shown
Title
Unlimiformer: Long-Range Transformers with Unlimited Length Input
Amanda Bertsch
Uri Alon
Graham Neubig
Matthew R. Gormley
RALM
94
122
0
02 May 2023
Training Language Models with Memory Augmentation
Zexuan Zhong
Tao Lei
Danqi Chen
RALM
226
126
0
25 May 2022
Big Bird: Transformers for Longer Sequences
Manzil Zaheer
Guru Guruganesh
Kumar Avinava Dubey
Joshua Ainslie
Chris Alberti
...
Philip Pham
Anirudh Ravula
Qifan Wang
Li Yang
Amr Ahmed
VLM
249
1,982
0
28 Jul 2020
Efficient Content-Based Sparse Attention with Routing Transformers
Aurko Roy
M. Saffar
Ashish Vaswani
David Grangier
MoE
234
578
0
12 Mar 2020
1