71
72

Near-Optimal Methods for Minimizing Star-Convex Functions and Beyond

Abstract

In this paper, we provide near-optimal accelerated first-order methods for minimizing a broad class of smooth nonconvex functions that are strictly unimodal on all lines through a minimizer. This function class, which we call the class of smooth quasar-convex functions, is parameterized by a constant γ(0,1]\gamma \in (0,1], where γ=1\gamma = 1 encompasses the classes of smooth convex and star-convex functions, and smaller values of γ\gamma indicate that the function can be "more nonconvex." We develop a variant of accelerated gradient descent that computes an ϵ\epsilon-approximate minimizer of a smooth γ\gamma-quasar-convex function with at most O(γ1ϵ1/2log(γ1ϵ1))O(\gamma^{-1} \epsilon^{-1/2} \log(\gamma^{-1} \epsilon^{-1})) total function and gradient evaluations. We also derive a lower bound of Ω(γ1ϵ1/2)\Omega(\gamma^{-1} \epsilon^{-1/2}) on the number of gradient evaluations required by any deterministic first-order method in the worst case, showing that, up to a logarithmic factor, no deterministic first-order algorithm can improve upon ours.

View on arXiv
Comments on this paper