A random matrix approach to neural networks

Citation
Louart, Cosme et al., A random matrix approach to neural networks, Annals of applied probability , 28(2), 2018, pp. 1190-1248
ISSN journal
10505164
Volume
28
Issue
2
Year of publication
2018
Pages
1190 - 1248
Database
ACNP
SICI code
Abstract
This article studies the Gram random matrix model G=1T.T., .=.(WX), classically found in the analysis of random feature maps and random neural networks, where X=[x1,.,xT].Rp.T is a (data) matrix of bounded norm, W.Rn.p is a matrix of independent zero-mean unit variance entries and .:R.R is a Lipschitz continuous (activation) function..(WX) being understood entry-wise. By means of a key concentration of measure lemma arising from nonasymptotic random matrix arguments, we prove that, as n,p,T grow large at the same rate, the resolvent Q=(G+.IT).1, for .>0, has a similar behavior as that met in sample covariance matrix models, involving notably the moment .=TnE[G], which provides in passing a deterministic equivalent for the empirical spectral measure of G. Application-wise, this result enables the estimation of the asymptotic performance of single-layer random neural networks. This in turn provides practical insights into the underlying mechanisms into play in random neural networks, entailing several unexpected consequences, as well as a fast practical means to tune the network hyperparameters.