24
6
v1v2 (latest)

Higher-Order Newton Methods with Polynomial Work per Iteration

Abstract

We present generalizations of Newton's method that incorporate derivatives of an arbitrary order dd but maintain a polynomial dependence on dimension in their cost per iteration. At each step, our dthd^{\text{th}}-order method uses semidefinite programming to construct and minimize a sum of squares-convex approximation to the dthd^{\text{th}}-order Taylor expansion of the function we wish to minimize. We prove that our dthd^{\text{th}}-order method has local convergence of order dd. This results in lower oracle complexity compared to the classical Newton method. We show on numerical examples that basins of attraction around local minima can get larger as dd increases. Under additional assumptions, we present a modified algorithm, again with polynomial cost per iteration, which is globally convergent and has local convergence of order dd.

View on arXiv
Comments on this paper