499

An O(sr)O(s^r)-Resolution ODE Framework for Discrete-Time Optimization Algorithms and Applications to the Linear Convergence of Saddle-Point Problems

Mathematical programming (Math. Program.), 2020
Abstract

There has been a long history of using Ordinary Differential Equations (ODEs) to understand the dynamic of discrete-time optimization algorithms. However, one major difficulty of applying this approach is that there can be multiple ODEs that correspond to the same discrete-time algorithm, depending on how to take the continuous limit, which makes it unclear how to obtain the suitable ODE from a discrete-time optimization algorithm. Inspired by the recent paper \cite{shi2018understanding}, we propose the rr-th degree ODE expansion of a discrete-time optimization algorithm, which provides a principal approach to construct the unique O(sr)O(s^r)-resolution ODE systems for a given discrete-time algorithm, where ss is the step-size of the algorithm. We utilize this machinery to study three classic algorithms -- gradient method (GM), proximal point method (PPM) and extra-gradient method (EGM) -- for finding the solution to the unconstrained convex-concave saddle-point problem minx\RRnmaxy\RRmL(x,y)\min_{x\in\RR^n} \max_{y\in \RR^m} L(x,y), which explains their puzzling convergent/divergent behaviors when L(x,y)L(x,y) is a bilinear function. Moreover, their O(s)O(s)-resolution ODEs inspire us to define the O(s)O(s)-linear-convergence condition on L(x,y)L(x,y), under which PPM and EGM exhabit linear convergence. This condition not only unifies the known linear convergence rate of PPM and EGM, but also showcases that these two algorithms exhibit linear convergence in broader contexts.

View on arXiv
Comments on this paper