Optimal Local Convergence Rates of Stochastic First-Order Methods under Local -PL
We study the local convergence rate of stochastic first-order methods under a local -Polyak-Lojasiewicz (-PL) condition in a neighborhood of a target connected component of the local minimizer set. The parameter is the exponent of the gradient norm in the -PL inequality: recovers the classical PL case, corresponds to Holder-type error bounds, and intermediate values interpolate between these regimes. Our performance criterion is the number of oracle queries required to output with , where for any . We work in a local regime where the algorithm is initialized near and, with high probability, its iterates remain in that neighborhood. We establish a lower bound for all stochastic first-order methods in this regime, and we obtain a matching upper bound for via a SARAH-type variance-reduced method with time-varying batch sizes and step sizes. In the convex setting, assuming a local -PL condition on the -sublevel set, we further show a complexity lower bound for reaching an -global optimum, matching the -dependence of known accelerated stochastic subgradient methods.
View on arXiv