Parameter Symmetry Breaking and Restoration Determines the Hierarchical Learning in AI Systems

The dynamics of learning in modern large AI systems is hierarchical, often characterized by abrupt, qualitative shifts akin to phase transitions observed in physical systems. While these phenomena hold promise for uncovering the mechanisms behind neural networks and language models, existing theories remain fragmented, addressing specific cases. In this paper, we posit that parameter symmetry breaking and restoration serve as a unifying mechanism underlying these behaviors. We synthesize prior observations and show how this mechanism explains three distinct hierarchies in neural networks: learning dynamics, model complexity, and representation formation. By connecting these hierarchies, we highlight symmetry -- a cornerstone of theoretical physics -- as a potential fundamental principle in modern AI.
View on arXiv@article{ziyin2025_2502.05300, title={ Parameter Symmetry Breaking and Restoration Determines the Hierarchical Learning in AI Systems }, author={ Liu Ziyin and Yizhou Xu and Tomaso Poggio and Isaac Chuang }, journal={arXiv preprint arXiv:2502.05300}, year={ 2025 } }