Analysis of Schedule-Free Nonconvex Optimization
First-order methods underpin most large-scale learning algorithms, yet their classical convergence guarantees hinge on carefully scheduled step-sizes that depend on the total horizon , which is rarely known in advance. The Schedule-Free (SF) method promises optimal performance with hyperparameters that are independent of by interpolating between Polyak--Ruppert averaging and momentum, but nonconvex analysis of SF has been limited or reliant on strong global assumptions. We introduce a robust Lyapunov framework that, under only -smoothness and lower-boundedness, reduces SF analysis to a single-step descent inequality. This yields horizon-agnostic bounds in the nonconvex setting: for constant step + PR averaging, for a linearly growing step-size, and a continuum of rates for polynomial averaging. We complement these proofs with Performance Estimation Problem (PEP) experiments that numerically validate our rates and suggest that our bound on the original nonconvex SF algorithm may tighten to . Our work extends SF's horizon-free guarantees to smooth nonconvex optimization and charts future directions for optimal nonconvex rates.
View on arXiv