Structural Risk Minimization for Regression
One means of fitting functions to high-dimensional data is by providing smoothness constraints. Recently, the following smooth function approximation problem was proposed by \citet*{herbert2014computing}: given a finite set and a function , interpolate the given information with a function (the class of first-order differentiable functions with Lipschitz gradients) such that for all , and the value of is minimal. An algorithm is provided that constructs such an approximating function and estimates the optimal Lipschitz constant in the noiseless setting. We address statistical aspects of reconstructing the approximating function from a closely-related class given samples from noisy data. We observe independent and identically distributed samples for , where is a noise term and the set is fixed and known. We obtain uniform bounds relating the empirical risk and true risk over the class , where the quantity grows with the number of samples at a rate governed by the metric entropy of the class . Finally, we provide an implementation using Vaidya's algorithm, supporting our results via numerical experiments on simulated data.
View on arXiv