ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2301.10945
8
67

A Fully First-Order Method for Stochastic Bilevel Optimization

26 January 2023
Jeongyeol Kwon
Dohyun Kwon
S. Wright
Robert D. Nowak
ArXivPDFHTML
Abstract

We consider stochastic unconstrained bilevel optimization problems when only the first-order gradient oracles are available. While numerous optimization methods have been proposed for tackling bilevel problems, existing methods either tend to require possibly expensive calculations regarding Hessians of lower-level objectives, or lack rigorous finite-time performance guarantees. In this work, we propose a Fully First-order Stochastic Approximation (F2SA) method, and study its non-asymptotic convergence properties. Specifically, we show that F2SA converges to an ϵ\epsilonϵ-stationary solution of the bilevel problem after ϵ−7/2,ϵ−5/2\epsilon^{-7/2}, \epsilon^{-5/2}ϵ−7/2,ϵ−5/2, and ϵ−3/2\epsilon^{-3/2}ϵ−3/2 iterations (each iteration using O(1)O(1)O(1) samples) when stochastic noises are in both level objectives, only in the upper-level objective, and not present (deterministic settings), respectively. We further show that if we employ momentum-assisted gradient estimators, the iteration complexities can be improved to ϵ−5/2,ϵ−4/2\epsilon^{-5/2}, \epsilon^{-4/2}ϵ−5/2,ϵ−4/2, and ϵ−3/2\epsilon^{-3/2}ϵ−3/2, respectively. We demonstrate even superior practical performance of the proposed method over existing second-order based approaches on MNIST data-hypercleaning experiments.

View on arXiv
Comments on this paper