ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2506.10378
111
1

Discovering Hierarchical Latent Capabilities of Language Models via Causal Representation Learning

12 June 2025
Jikai Jin
Vasilis Syrgkanis
Sham Kakade
Hanlin Zhang
    ELM
ArXiv (abs)PDFHTML
Main:11 Pages
36 Figures
Bibliography:7 Pages
4 Tables
Appendix:24 Pages
Abstract

Faithful evaluation of language model capabilities is crucial for deriving actionable insights that can inform model development. However, rigorous causal evaluations in this domain face significant methodological challenges, including complex confounding effects and prohibitive computational costs associated with extensive retraining. To tackle these challenges, we propose a causal representation learning framework wherein observed benchmark performance is modeled as a linear transformation of a few latent capability factors. Crucially, these latent factors are identified as causally interrelated after appropriately controlling for the base model as a common confounder. Applying this approach to a comprehensive dataset encompassing over 1500 models evaluated across six benchmarks from the Open LLM Leaderboard, we identify a concise three-node linear causal structure that reliably explains the observed performance variations. Further interpretation of this causal structure provides substantial scientific insights beyond simple numerical rankings: specifically, we reveal a clear causal direction starting from general problem-solving capabilities, advancing through instruction-following proficiency, and culminating in mathematical reasoning ability. Our results underscore the essential role of carefully controlling base model variations during evaluation, a step critical to accurately uncovering the underlying causal relationships among latent model capabilities.

View on arXiv
@article{jin2025_2506.10378,
  title={ Discovering Hierarchical Latent Capabilities of Language Models via Causal Representation Learning },
  author={ Jikai Jin and Vasilis Syrgkanis and Sham Kakade and Hanlin Zhang },
  journal={arXiv preprint arXiv:2506.10378},
  year={ 2025 }
}
Comments on this paper