ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2209.08951
31
9

Generalization Bounds for Stochastic Gradient Descent via Localized ε\varepsilonε-Covers

19 September 2022
Sejun Park
Umut Simsekli
Murat A. Erdogdu
ArXivPDFHTML
Abstract

In this paper, we propose a new covering technique localized for the trajectories of SGD. This localization provides an algorithm-specific complexity measured by the covering number, which can have dimension-independent cardinality in contrast to standard uniform covering arguments that result in exponential dimension dependency. Based on this localized construction, we show that if the objective function is a finite perturbation of a piecewise strongly convex and smooth function with PPP pieces, i.e. non-convex and non-smooth in general, the generalization error can be upper bounded by O((log⁡nlog⁡(nP))/n)O(\sqrt{(\log n\log(nP))/n})O((lognlog(nP))/n​), where nnn is the number of data samples. In particular, this rate is independent of dimension and does not require early stopping and decaying step size. Finally, we employ these results in various contexts and derive generalization bounds for multi-index linear models, multi-class support vector machines, and KKK-means clustering for both hard and soft label setups, improving the known state-of-the-art rates.

View on arXiv
Comments on this paper