Conditional regression for the Nonlinear Single-Variable Model
Several statistical models for regression of a function on without the statistical and computational curse of dimensionality exist, for example by imposing and exploiting geometric assumptions on the distribution of the data (e.g. that its support is low-dimensional), or strong smoothness assumptions on , or a special structure . Among the latter, compositional models assume with mapping to with , have been studied, and include classical single- and multi-index models and recent works on neural networks. While the case where is linear is rather well-understood, much less is known when is nonlinear, and in particular for which 's the curse of dimensionality in estimating , or both and , may be circumvented. In this paper, we consider a model $F(X):=f(\Pi_\gamma X) $ where is the closest-point projection onto the parameter of a regular curve and . The input data is not low-dimensional, far from , conditioned on being well-defined. The distribution of the data, and are unknown. This model is a natural nonlinear generalization of the single-index model, which corresponds to being a line. We propose a nonparametric estimator, based on conditional regression, and show that under suitable assumptions, the strongest of which being that is coarsely monotone, it can achieve the - optimal min-max rate for non-parametric regression, up to the level of noise in the observations, and be constructed in time . All the constants in the learning bounds, in the minimal number of samples required for our bounds to hold, and in the computational complexity are at most low-order polynomials in .
View on arXiv