11
7

Greedy Adversarial Equilibrium: An Efficient Alternative to Nonconvex-Nonconcave Min-Max Optimization

Abstract

Min-max optimization of an objective function f:Rd×RdRf: \mathbb{R}^d \times \mathbb{R}^d \rightarrow \mathbb{R} is an important model for robustness in an adversarial setting, with applications to many areas including optimization, economics, and deep learning. In many applications ff may be nonconvex-nonconcave, and finding a global min-max point may be computationally intractable. There is a long line of work that seeks computationally tractable algorithms for alternatives to the min-max optimization model. However, many of the alternative models have solution points which are only guaranteed to exist under strong assumptions on ff, such as convexity, monotonicity, or special properties of the starting point. We propose an optimization model, the ε\varepsilon-greedy adversarial equilibrium, and show that it can serve as a computationally tractable alternative to the min-max optimization model. Roughly, we say that a point (x,y)(x^\star, y^\star) is an ε\varepsilon-greedy adversarial equilibrium if yy^\star is an ε\varepsilon-approximate local maximum for f(x,)f(x^\star,\cdot), and xx^\star is an ε\varepsilon-approximate local minimum for a "greedy approximation" to the function maxzf(x,z)\max_z f(x, z) which can be efficiently estimated using second-order optimization algorithms. We prove the existence of such a point for any smooth function which is bounded and has Lipschitz Hessian. To prove existence, we introduce an algorithm that converges from any starting point to an ε\varepsilon-greedy adversarial equilibrium in a number of evaluations of the function ff, the max-player's gradient yf(x,y)\nabla_y f(x,y), and its Hessian y2f(x,y)\nabla^2_y f(x,y), that is polynomial in the dimension dd, 1/ε1/\varepsilon, and the bounds on ff and its Lipschitz constant.

View on arXiv
Comments on this paper