730

Optimal Computational and Statistical Rates of Convergence for Sparse Nonconvex Learning Problems

Annals of Statistics (AoS), 2013
Abstract

We provide theoretical analysis of the statistical and computational properties of penalized MM-estimators that can be formulated as the solution to a possibly nonconvex optimization problem. Many important estimators fall in this category, including least squares regression with nonconvex regularization, generalized linear models with nonconvex regularization, and sparse elliptical random design regression. For these problems it is intractable to calculate the global solution due to the nonconvex formulation. In this paper, we propose an approximate regularization path following algorithm for solving a variety of learning problems with nonconvex objective functions. Under a unified analytical framework, we simultaneously provide explicit statistical and computational rates of convergence of arbitrary local solution obtained by the algorithm. Computationally, our algorithm attains a global geometric rate of convergence for calculating the full regularization path, which is optimal among all first-order algorithms. Unlike most existing methods which only attain geometric rates of convergence for one single regularization parameter, our algorithm calculates the full regularization path with the same iteration complexity. In particular, we provide a refined iteration complexity bound to sharply characterize the performance of each stage along the regularization path. Statistically, we provide sharp sample complexity analysis for all the approximate local solutions along the regularization path. In particular, our analysis improves upon existing results by showing a more refined sample complexity bound for the final estimator. This result shows that the final estimator attains an oracle statistical property due to the usage of nonconvex penalty. Thorough numerical results are provided to back up our theoretical analysis.

View on arXiv
Comments on this paper