ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2412.00059
89
1

A Learn-to-Optimize Approach for Coordinate-Wise Step Sizes for Quasi-Newton Methods

25 November 2024
Wei Lin
Qingyu Song
Hong Xu
ArXivPDFHTML
Abstract

Tuning step sizes is crucial for the stability and efficiency of optimization algorithms. While adaptive coordinate-wise step sizes have been shown to outperform scalar step size in first-order methods, their use in second-order methods is still under-explored and more challenging. Current approaches, including hypergradient descent and cutting plane methods, offer limited improvements or encounter difficulties in second-order contexts. To address these limitations, we first conduct a theoretical analysis within the Broyden-Fletcher-Goldfarb-Shanno (BFGS) framework, a prominent quasi-Newton method, and derive sufficient conditions for coordinate-wise step sizes that ensure convergence and stability. Building on this theoretical foundation, we introduce a novel learn-to-optimize (L2O) method that employs LSTM-based networks to learn optimal step sizes by leveraging insights from past optimization trajectories, while inherently respecting the derived theoretical guarantees. Extensive experiments demonstrate that our approach achieves substantial improvements over scalar step size methods and hypergradient descent-based method, offering up to 4×\times× faster convergence across diverse optimization tasks.

View on arXiv
@article{lin2025_2412.00059,
  title={ A Learn-to-Optimize Approach for Coordinate-Wise Step Sizes for Quasi-Newton Methods },
  author={ Wei Lin and Qingyu Song and Hong Xu },
  journal={arXiv preprint arXiv:2412.00059},
  year={ 2025 }
}
Comments on this paper