ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2403.07264
25
2

Near-Interpolators: Rapid Norm Growth and the Trade-Off between Interpolation and Generalization

12 March 2024
Yutong Wang
Rishi Sonthalia
Wei Hu
ArXivPDFHTML
Abstract

We study the generalization capability of nearly-interpolating linear regressors: β\boldsymbol{\beta}β's whose training error τ\tauτ is positive but small, i.e., below the noise floor. Under a random matrix theoretic assumption on the data distribution and an eigendecay assumption on the data covariance matrix Σ\boldsymbol{\Sigma}Σ, we demonstrate that any near-interpolator exhibits rapid norm growth: for τ\tauτ fixed, β\boldsymbol{\beta}β has squared ℓ2\ell_2ℓ2​-norm E[∥β∥22]=Ω(nα)\mathbb{E}[\|{\boldsymbol{\beta}}\|_{2}^{2}] = \Omega(n^{\alpha})E[∥β∥22​]=Ω(nα) where nnn is the number of samples and α>1\alpha >1α>1 is the exponent of the eigendecay, i.e., λi(Σ)∼i−α\lambda_i(\boldsymbol{\Sigma}) \sim i^{-\alpha}λi​(Σ)∼i−α. This implies that existing data-independent norm-based bounds are necessarily loose. On the other hand, in the same regime we precisely characterize the asymptotic trade-off between interpolation and generalization. Our characterization reveals that larger norm scaling exponents α\alphaα correspond to worse trade-offs between interpolation and generalization. We verify empirically that a similar phenomenon holds for nearly-interpolating shallow neural networks.

View on arXiv
Comments on this paper