41
5

Anderson Acceleration as a Krylov Method with Application to Asymptotic Convergence Analysis

Abstract

Anderson acceleration is widely used for accelerating the convergence of fixed-point methods xk+1=q(xk)x_{k+1}=q(x_{k}), xkRnx_k \in \mathbb{R}^n. We consider the case of linear fixed-point methods xk+1=Mxk+bx_{k+1}=M x_{k}+b and obtain polynomial residual update formulas for AA(mm), i.e., Anderson acceleration with window size mm. We find that the standard AA(mm) method with initial iterates xkx_k, k=0,,mk=0, \ldots, m defined recursively using AA(kk), is a Krylov space method. This immediately implies that kk iterations of AA(mm) cannot produce a smaller residual than kk iterations of GMRES without restart (but without implying anything about the relative convergence speed of (windowed) AA(mm) versus restarted GMRES(mm)). We introduce the notion of multi-Krylov method and show that AA(mm) with general initial iterates {x0,,xm}\{x_0, \ldots, x_m\} is a multi-Krylov method. We find that the AA(mm) residual polynomials observe a periodic memory effect where increasing powers of the error iteration matrix MM act on the initial residual as the iteration number increases. We derive several further results based on these polynomial residual update formulas, including orthogonality relations, a lower bound on the AA(1) acceleration coefficient βk\beta_k, and explicit nonlinear recursions for the AA(1) residuals and residual polynomials that do not include the acceleration coefficient βk\beta_k. We apply these results to study the influence of the initial guess on the asymptotic convergence factor of AA(1).

View on arXiv
Comments on this paper