ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2105.06548
  4. Cited By
Not All Memories are Created Equal: Learning to Forget by Expiring
v1v2 (latest)

Not All Memories are Created Equal: Learning to Forget by Expiring

13 May 2021
Sainbayar Sukhbaatar
Da Ju
Spencer Poff
Stephen Roller
Arthur Szlam
Jason Weston
Angela Fan
    CLL
ArXiv (abs)PDFHTML

Papers citing "Not All Memories are Created Equal: Learning to Forget by Expiring"

14 / 14 papers shown
Title
Lifting Data-Tracing Machine Unlearning to Knowledge-Tracing for Foundation Models
Lifting Data-Tracing Machine Unlearning to Knowledge-Tracing for Foundation Models
Yuwen Tan
Boqing Gong
MU
119
0
0
12 Jun 2025
Multi-Token Attention
Multi-Token Attention
O. Yu. Golovneva
Tianlu Wang
Jason Weston
Sainbayar Sukhbaatar
134
2
0
01 Apr 2025
MemoNav: Working Memory Model for Visual Navigation
MemoNav: Working Memory Model for Visual Navigation
Hongxin Li
Zeyu Wang
Xueke Yang
Yu-Ren Yang
Shuqi Mei
Zhaoxiang Zhang
207
6
0
29 Feb 2024
A Data Source for Reasoning Embodied Agents
A Data Source for Reasoning Embodied Agents
Jack Lanchantin
Sainbayar Sukhbaatar
Gabriel Synnaeve
Yuxuan Sun
Kavya Srinet
Arthur Szlam
LM&RoLRM
92
7
0
14 Sep 2023
Memory in humans and deep language models: Linking hypotheses for model
  augmentation
Memory in humans and deep language models: Linking hypotheses for model augmentation
Omri Raccah
Pheobe Chen
Ted Willke
David Poeppel
Vy A. Vo
RALM
122
1
0
04 Oct 2022
Training Language Models with Memory Augmentation
Training Language Models with Memory Augmentation
Zexuan Zhong
Tao Lei
Danqi Chen
RALM
490
138
0
25 May 2022
Memorizing Transformers
Memorizing Transformers
Yuhuai Wu
M. Rabe
DeLesley S. Hutchins
Christian Szegedy
RALM
148
194
0
16 Mar 2022
Block-Recurrent Transformers
Block-Recurrent Transformers
DeLesley S. Hutchins
Imanol Schlag
Yuhuai Wu
Ethan Dyer
Behnam Neyshabur
201
113
0
11 Mar 2022
MeMViT: Memory-Augmented Multiscale Vision Transformer for Efficient
  Long-Term Video Recognition
MeMViT: Memory-Augmented Multiscale Vision Transformer for Efficient Long-Term Video Recognition
Chao-Yuan Wu
Yanghao Li
K. Mangalam
Haoqi Fan
Bo Xiong
Jitendra Malik
Christoph Feichtenhofer
ViT
240
221
0
20 Jan 2022
Learning what to remember
Learning what to remember
Robi Bhattacharjee
G. Mahajan
CLLKELM
61
0
0
11 Jan 2022
Discourse-Aware Soft Prompting for Text Generation
Discourse-Aware Soft Prompting for Text Generation
Marjan Ghazvininejad
Vladimir Karpukhin
Vera Gor
Asli Celikyilmaz
97
7
0
10 Dec 2021
Hierarchical Transformers Are More Efficient Language Models
Hierarchical Transformers Are More Efficient Language Models
Piotr Nawrot
Szymon Tworkowski
Michał Tyrolski
Lukasz Kaiser
Yuhuai Wu
Christian Szegedy
Henryk Michalewski
184
82
0
26 Oct 2021
ABC: Attention with Bounded-memory Control
ABC: Attention with Bounded-memory Control
Hao Peng
Jungo Kasai
Nikolaos Pappas
Dani Yogatama
Zhaofeng Wu
Lingpeng Kong
Roy Schwartz
Noah A. Smith
172
27
0
06 Oct 2021
BERTnesia: Investigating the capture and forgetting of knowledge in BERT
BERTnesia: Investigating the capture and forgetting of knowledge in BERT
Jonas Wallat
Jaspreet Singh
Avishek Anand
CLLKELM
222
62
0
05 Jun 2021
1