On the Randomized Complexity of Minimizing a Convex Quadratic Function

Minimizing a convex, quadratic objective of the form for is a fundamental problem in machine learning and optimization. In this work, we prove gradient-query complexity lower bounds for minimizing convex quadratic functions which apply to both deterministic and \emph{randomized} algorithms. Specifically, for , we exhibit a distribution over with condition number , such that any \emph{randomized} algorithm requires gradient queries to find a solution for which , where is the optimal solution, and a small constant. Setting , this lower bound implies the minimax rate of queries required to minimize an arbitrary convex quadratic function up to error . Our lower bound holds for a distribution derived from classical ensembles in random matrix theory, and relies on a careful reduction from adaptively estimating a planted vector in a deformed Wigner model. A key step in deriving sharp lower bounds is demonstrating that the optimization error cannot align too closely with . To this end, we prove an upper bound on the cosine between and in terms of the MMSE of estimating the plant in a deformed Wigner model. We then bound the MMSE by carefully modifying a result due to Lelarge and Miolane 2016, which rigorously establishes a general replica-symmetric formula for planted matrix models.
View on arXiv