ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2509.17196
96
0

Evolution of Concepts in Language Model Pre-Training

21 September 2025
Xuyang Ge
Wentao Shu
Jiaxing Wu
Yunhua Zhou
Zhengfu He
Xipeng Qiu
ArXiv (abs)PDFHTMLHuggingFace (1 upvotes)
Main:10 Pages
26 Figures
Bibliography:2 Pages
1 Tables
Appendix:18 Pages
Abstract

Language models obtain extensive capabilities through pre-training. However, the pre-training process remains a black box. In this work, we track linear interpretable feature evolution across pre-training snapshots using a sparse dictionary learning method called crosscoders. We find that most features begin to form around a specific point, while more complex patterns emerge in later training stages. Feature attribution analyses reveal causal connections between feature evolution and downstream performance. Our feature-level observations are highly consistent with previous findings on Transformer's two-stage learning process, which we term a statistical learning phase and a feature learning phase. Our work opens up the possibility to track fine-grained representation progress during language model learning dynamics.

View on arXiv
Comments on this paper