89
v1v2 (latest)

Structural Risk Minimization for C1,1(Rd)C^{1,1}(\mathbb{R}^d) Regression

Abstract

One means of fitting functions to high-dimensional data is by providing smoothness constraints. Recently, the following smooth function approximation problem was proposed: given a finite set ERdE \subset \mathbb{R}^d and a function f:ERf: E \rightarrow \mathbb{R}, interpolate the given information with a function f^C˙1,1(Rd)\widehat{f} \in \dot{C}^{1, 1}(\mathbb{R}^d) (the class of first-order differentiable functions with Lipschitz gradients) such that f^(a)=f(a)\widehat{f}(a) = f(a) for all aEa \in E, and the value of Lip(f^)\mathrm{Lip}(\nabla \widehat{f}) is minimal. An algorithm is provided that constructs such an approximating function f^\widehat{f} and estimates the optimal Lipschitz constant Lip(f^)\mathrm{Lip}(\nabla \widehat{f}) in the noiseless setting. We address statistical aspects of reconstructing the approximating function f^\widehat{f} from a closely-related class C1,1(Rd)C^{1, 1}(\mathbb{R}^d) given samples from noisy data. We observe independent and identically distributed samples y(a)=f(a)+ξ(a)y(a) = f(a) + \xi(a) for aEa \in E, where ξ(a)\xi(a) is a noise term and the set ERdE \subset \mathbb{R}^d is fixed and known. We obtain uniform bounds relating the empirical risk and true risk over the class FM~={fC1,1(Rd)Lip(f)M~}\mathcal{F}_{\widetilde{M}} = \{f \in C^{1, 1}(\mathbb{R}^d) \mid \mathrm{Lip}(\nabla f) \leq \widetilde{M}\}, where the quantity M~\widetilde{M} grows with the number of samples at a rate governed by the metric entropy of the class C1,1(Rd)C^{1, 1}(\mathbb{R}^d). Finally, we provide an implementation using Vaidya's algorithm, supporting our results via numerical experiments on simulated data.

View on arXiv
Comments on this paper