440
v1v2v3 (latest)

Structural Inference: Interpreting Small Language Models with Susceptibilities

Main:9 Pages
40 Figures
Bibliography:4 Pages
3 Tables
Appendix:39 Pages
Abstract

We develop a linear response framework for interpretability that treats a neural network as a Bayesian statistical mechanical system. A small perturbation of the data distribution, for example shifting the Pile toward GitHub or legal text, induces a first-order change in the posterior expectation of an observable localized on a chosen component of the network. The resulting susceptibility can be estimated efficiently with local SGLD samples and factorizes into signed, per-token contributions that serve as attribution scores. We combine these susceptibilities into a response matrix whose low-rank structure separates functional modules such as multigram and induction heads in a 3M-parameter transformer.

View on arXiv
Comments on this paper