An -Resolution ODE Framework for Discrete-Time Optimization
Algorithms and Applications to the Linear Convergence of Minimax Problems
There has been a long history of using Ordinary Differential Equations (ODEs) to understand the dynamic of discrete-time algorithms (DTAs). However, there are two major difficulties to apply this approach: (i) it is unclear how to obtain a suitable ODE from a DTA, and (ii) it is unclear what is the connection between the convergence of a DTA and the convergence of its corresponding ODE. Inspired by the recent work \cite{shi2018understanding}, we propose an -resolution ODE framework, which (partially) resolves the above two difficulties. More specifically, we propose the -th degree ODE expansion of a discrete-time optimization algorithm, which provides a principal approach to construct the unique -resolution ODE for a given DTA, where is the step-size of the algorithm. Furthermore, we propose the -linear-convergence condition of a DTA under which the -resolution ODE converges linearly to optimal solution. These conditions are usually obvious from the -resolution ODE, and more importantly, we show that such conditions can automatically guarantee the linear convergence of a large class of DTAs. To better illustrate this machinery, we utilize it to study three classic algorithms -- gradient method (GM), proximal point method (PPM) and extra-gradient method (EGM) -- for solving the unconstrained minimax problem . Their -resolution ODEs explain the puzzling convergent/divergent behaviors of GM, PPM and EGM when is a bilinear function. Moreover, the -linear-convergence condition on not only unifies the known linear convergence rate of PPM and EGM, but also showcases that these two algorithms exhibit linear convergence in broader contexts, including solving a class of nonconvex-nonconcave minimax problems.
View on arXiv