ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2505.15962
336
0
v1v2v3 (latest)

Pre-training Large Memory Language Models with Internal and External Knowledge

21 May 2025
Linxi Zhao
Sofian Zalouk
Christian K. Belardi
Justin Lovelace
Jin Peng Zhou
Ryan Thomas Noonan
Dongyoung Go
Kilian Q. Weinberger
    KELMHILM
ArXiv (abs)PDFHTMLGithub (16★)
Main:10 Pages
13 Figures
Bibliography:5 Pages
22 Tables
Appendix:16 Pages
Abstract

Neural language models are black-boxes -- both linguistic patterns and factual knowledge are distributed across billions of opaque parameters. This entangled encoding makes it difficult to reliably inspect, verify, or update specific facts. We propose a new class of language models, Large Memory Language Models (LMLM) with a pre-training recipe that stores factual knowledge in both internal weights and an external database. Our approach strategically masks externally retrieved factual values from the training loss, thereby teaching the model to perform targeted lookups rather than relying on memorization in model weights. Our experiments demonstrate that LMLMs achieve competitive performance compared to significantly larger, knowledge-dense LLMs on standard benchmarks, while offering the advantages of explicit, editable, and verifiable knowledge bases. This work represents a fundamental shift in how language models interact with and manage factual knowledge.

View on arXiv
Comments on this paper