22
2

Bounds for the smallest eigenvalue of the NTK for arbitrary spherical data of arbitrary dimension

Abstract

Bounds on the smallest eigenvalue of the neural tangent kernel (NTK) are a key ingredient in the analysis of neural network optimization and memorization. However, existing results require distributional assumptions on the data and are limited to a high-dimensional setting, where the input dimension d0d_0 scales at least logarithmically in the number of samples nn. In this work we remove both of these requirements and instead provide bounds in terms of a measure of the collinearity of the data: notably these bounds hold with high probability even when d0d_0 is held constant versus nn. We prove our results through a novel application of the hemisphere transform.

View on arXiv
Comments on this paper