Learning Orthogonal Multi-Index Models: A Fine-Grained Information Exponent Analysis
The information exponent ([BAGJ21]) and its extensions -- which are equivalent to the lowestdegree in the Hermite expansion of the link function (after a potential label transform) for Gaussian single-indexmodels -- have played an important role in predicting the sample complexity of online stochastic gradient descent(SGD) in various learning tasks. In this work, we demonstrate that, for multi-index models, focusing solely on thelowest degree can miss key structural details of the model and result in suboptimal rates.Specifically, we consider the task of learning target functions of form ,where , the ground-truth directions are orthonormal, and the information exponent of is . Based on the theory of information exponent, when , only the relevant subspace (not the exactdirections) can be recovered due to the rotational invariance of the second-order terms, and when ,recovering the directions using online SGD require samples. In this work, we show that byconsidering both second- and higher-order terms, we can first learn the relevant space using the second-orderterms, and then the exact directions using the higher-order terms, and the overall sample and complexity of onlineSGD is .
View on arXiv