ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2507.09823
132
1
v1v2 (latest)

Nesterov Finds GRAAL: Optimal and Adaptive Gradient Method for Convex Optimization

13 July 2025
Ekaterina Borodich
D. Kovalev
ArXiv (abs)PDFHTML
Main:9 Pages
Bibliography:3 Pages
1 Tables
Appendix:14 Pages
Abstract

In this paper, we focus on the problem of minimizing a continuously differentiable convex objective function, min⁡xf(x)\min_x f(x)minx​f(x). Recently, Malitsky (2020); Alacaoglu et al.(2023) developed an adaptive first-order method, GRAAL. This algorithm computes stepsizes by estimating the local curvature of the objective function without any line search procedures or hyperparameter tuning, and attains the standard iteration complexity O(L∥x0−x∗∥2/ϵ)\mathcal{O}(L\lVert x_0-x^*\rVert^2/\epsilon)O(L∥x0​−x∗∥2/ϵ) of fixed-stepsize gradient descent for LLL-smooth functions. However, a natural question arises: is it possible to accelerate the convergence of GRAAL to match the optimal complexity O(L∥x0−x∗∥2/ϵ)\mathcal{O}(\sqrt{L\lVert x_0-x^*\rVert^2/\epsilon})O(L∥x0​−x∗∥2/ϵ​) of the accelerated gradient descent of Nesterov (1983)? Although some attempts have been made by Li and Lan (2025); Suh and Ma (2025), the ability of existing accelerated algorithms to adapt to the local curvature of the objective function is highly limited. We resolve this issue and develop GRAAL with Nesterov acceleration, which can adapt its stepsize to the local curvature at a geometric, or linear, rate just like non-accelerated GRAAL. We demonstrate the adaptive capabilities of our algorithm by proving that it achieves near-optimal iteration complexities for LLL-smooth functions, as well as under a more general (L0,L1)(L_0,L_1)(L0​,L1​)-smoothness assumption (Zhang et al., 2019).

View on arXiv
Comments on this paper