ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1702.05419
54
161

A Random Matrix Approach to Neural Networks

17 February 2017
Cosme Louart
Zhenyu Liao
Romain Couillet
ArXivPDFHTML
Abstract

This article studies the Gram random matrix model G=1TΣTΣG=\frac1T\Sigma^{\rm T}\SigmaG=T1​ΣTΣ, Σ=σ(WX)\Sigma=\sigma(WX)Σ=σ(WX), classically found in the analysis of random feature maps and random neural networks, where X=[x1,…,xT]∈Rp×TX=[x_1,\ldots,x_T]\in{\mathbb R}^{p\times T}X=[x1​,…,xT​]∈Rp×T is a (data) matrix of bounded norm, W∈Rn×pW\in{\mathbb R}^{n\times p}W∈Rn×p is a matrix of independent zero-mean unit variance entries, and σ:R→R\sigma:{\mathbb R}\to{\mathbb R}σ:R→R is a Lipschitz continuous (activation) function --- σ(WX)\sigma(WX)σ(WX) being understood entry-wise. By means of a key concentration of measure lemma arising from non-asymptotic random matrix arguments, we prove that, as n,p,Tn,p,Tn,p,T grow large at the same rate, the resolvent Q=(G+γIT)−1Q=(G+\gamma I_T)^{-1}Q=(G+γIT​)−1, for γ>0\gamma>0γ>0, has a similar behavior as that met in sample covariance matrix models, involving notably the moment Φ=TnE[G]\Phi=\frac{T}n{\mathbb E}[G]Φ=nT​E[G], which provides in passing a deterministic equivalent for the empirical spectral measure of GGG. Application-wise, this result enables the estimation of the asymptotic performance of single-layer random neural networks. This in turn provides practical insights into the underlying mechanisms into play in random neural networks, entailing several unexpected consequences, as well as a fast practical means to tune the network hyperparameters.

View on arXiv
Comments on this paper