Extensions of Smoothing via Taut Strings
Suppose that we observe independent random pairs (X_1,Y_1), (X_2,Y_2), ..., (X_n,Y_n).Our goal is to estimate regression functions such as the conditional mean or beta-quantile of Y given X, where 0 < beta < 1. In order to achieve this we minimize criteria such as, for instance, the sum of rho(f(X_i) - Y_i) over all i plus lambda * TV(f) among all candidate functions. Here rho(.) is some convex function depending on the particular regression function we have in mind, TV(f) stands for the total variation of f, and lambda > 0 is some tuning parameter. This framework is extended further to include binary or Poisson regression, and to include localized total variation penalties. The latter are needed to construct estimators adapting to inhomogeneous smoothness of f. For the general framework we develop noniterative algorithms for the solution of the minimization problems which are closely related to the taut string algorithm (cf. Davies and Kovac, 2001). Further we establish a connection between the present setting and monotone regression, extending previous work by Mammen and van de Geer (1997).
View on arXiv