A Random Matrix Approach to Neural Networks
This article studies the Gram random matrix model , , classically found in random neural networks, where is a (data) matrix of bounded norm, is a matrix of independent zero-mean unit variance entries, and is a Lipschitz continuous (activation) function --- being understood entry-wise. We prove that, as grow large at the same rate, the resolvent , for , has a similar behavior as that met in sample covariance matrix models, involving notably the moment , which provides in passing a deterministic equivalent for the empirical spectral measure of . This result, established by means of concentration of measure arguments, enables the estimation of the asymptotic performance of single-layer random neural networks. This in turn provides practical insights into the underlying mechanisms into play in random neural networks, entailing several unexpected consequences, as well as a fast practical means to tune the network hyperparameters.
View on arXiv