ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2402.03009
  4. Cited By
UniMem: Towards a Unified View of Long-Context Large Language Models

UniMem: Towards a Unified View of Long-Context Large Language Models

5 February 2024
Junjie Fang
Likai Tang
Hongzhe Bi
Yujia Qin
Si Sun
Zhenyu Li
Haolun Li
Yongjian Li
Xin Cong
Yukun Yan
Xiaodong Shi
Sen Song
Yankai Lin
Zhiyuan Liu
Maosong Sun
ArXivPDFHTML

Papers citing "UniMem: Towards a Unified View of Long-Context Large Language Models"

4 / 4 papers shown
Title
Unlimiformer: Long-Range Transformers with Unlimited Length Input
Unlimiformer: Long-Range Transformers with Unlimited Length Input
Amanda Bertsch
Uri Alon
Graham Neubig
Matthew R. Gormley
RALM
94
122
0
02 May 2023
Training Language Models with Memory Augmentation
Training Language Models with Memory Augmentation
Zexuan Zhong
Tao Lei
Danqi Chen
RALM
232
126
0
25 May 2022
Big Bird: Transformers for Longer Sequences
Big Bird: Transformers for Longer Sequences
Manzil Zaheer
Guru Guruganesh
Kumar Avinava Dubey
Joshua Ainslie
Chris Alberti
...
Philip Pham
Anirudh Ravula
Qifan Wang
Li Yang
Amr Ahmed
VLM
249
1,982
0
28 Jul 2020
Efficient Content-Based Sparse Attention with Routing Transformers
Efficient Content-Based Sparse Attention with Routing Transformers
Aurko Roy
M. Saffar
Ashish Vaswani
David Grangier
MoE
238
578
0
12 Mar 2020
1