ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2505.14424
  4. Cited By
Explaining Neural Networks with Reasons

Explaining Neural Networks with Reasons

20 May 2025
Levin Hornischer
Hannes Leitgeb
    FAttAAMLMILM
ArXiv (abs)PDFHTML

Papers citing "Explaining Neural Networks with Reasons"

3 / 3 papers shown
Title
Sparse Autoencoders Can Interpret Randomly Initialized Transformers
Sparse Autoencoders Can Interpret Randomly Initialized Transformers
Thomas Heap
Tim Lawson
Lucy Farnik
Laurence Aitchison
81
17
0
29 Jan 2025
A Practical Review of Mechanistic Interpretability for Transformer-Based Language Models
A Practical Review of Mechanistic Interpretability for Transformer-Based Language Models
Daking Rai
Yilun Zhou
Shi Feng
Abulhair Saparov
Ziyu Yao
190
33
0
02 Jul 2024
Standards for Belief Representations in LLMs
Standards for Belief Representations in LLMs
Daniel A. Herrmann
B. Levinstein
99
11
0
31 May 2024
1