104
287

Minimax-optimal rates for sparse additive models over kernel classes via convex programming

Abstract

Sparse additive models are families of dd-variate functions that have the additive decomposition f=jSfjf^* = \sum_{j \in S} f^*_j, where SS is an unknown subset of cardinality sds \ll d. In this paper, we consider the case where each univariate component function fjf^*_j lies in a reproducing kernel Hilbert space (RKHS), and analyze a method for estimating the unknown function ff^* based on kernels combined with 1\ell_1-type convex regularization. Working within a high-dimensional framework that allows both the dimension dd and sparsity ss to increase with nn, we derive convergence rates (upper bounds) in the L2(P)L^2(\mathbb{P}) and L2(Pn)L^2(\mathbb{P}_n) norms over the class \MyBigClass\MyBigClass of sparse additive models with each univariate function fjf^*_j in the unit ball of a univariate RKHS with bounded kernel function. We complement our upper bounds by deriving minimax lower bounds on the L2(P)L^2(\mathbb{P}) error, thereby showing the optimality of our method. Thus, we obtain optimal minimax rates for many interesting classes of sparse additive models, including polynomials, splines, and Sobolev classes. We also show that if, in contrast to our univariate conditions, the multivariate function class is assumed to be globally bounded, then much faster estimation rates are possible for any sparsity s=Ω(n)s = \Omega(\sqrt{n}), showing that global boundedness is a significant restriction in the high-dimensional setting.

View on arXiv
Comments on this paper