31
53

Asymptotic behavior of p\ell_p-based Laplacian regularization in semi-supervised learning

Abstract

Given a weighted graph with NN vertices, consider a real-valued regression problem in a semi-supervised setting, where one observes nn labeled vertices, and the task is to label the remaining ones. We present a theoretical study of p\ell_p-based Laplacian regularization under a dd-dimensional geometric random graph model. We provide a variational characterization of the performance of this regularized learner as NN grows to infinity while nn stays constant, the associated optimality conditions lead to a partial differential equation that must be satisfied by the associated function estimate f^\hat{f}. From this formulation we derive several predictions on the limiting behavior the dd-dimensional function f^\hat{f}, including (a) a phase transition in its smoothness at the threshold p=d+1p = d + 1, and (b) a tradeoff between smoothness and sensitivity to the underlying unlabeled data distribution PP. Thus, over the range pdp \leq d, the function estimate f^\hat{f} is degenerate and "spiky," whereas for pd+1p\geq d+1, the function estimate f^\hat{f} is smooth. We show that the effect of the underlying density vanishes monotonically with pp, such that in the limit p=p = \infty, corresponding to the so-called Absolutely Minimal Lipschitz Extension, the estimate f^\hat{f} is independent of the distribution PP. Under the assumption of semi-supervised smoothness, ignoring PP can lead to poor statistical performance, in particular, we construct a specific example for d=1d=1 to demonstrate that p=2p=2 has lower risk than p=p=\infty due to the former penalty adapting to PP and the latter ignoring it. We also provide simulations that verify the accuracy of our predictions for finite sample sizes. Together, these properties show that p=d+1p = d+1 is an optimal choice, yielding a function estimate f^\hat{f} that is both smooth and non-degenerate, while remaining maximally sensitive to PP.

View on arXiv
Comments on this paper