Tree-Projected Gradient Descent for Estimating Gradient-Sparse Parameters on Graphs

We study estimation of a gradient-sparse parameter vector , having strong gradient-sparsity on an underlying graph . Given observations and a smooth, convex loss function for which minimizes the population risk , we propose to estimate by a projected gradient descent algorithm that iteratively and approximately projects gradient steps onto spaces of vectors having small gradient-sparsity over low-degree spanning trees of . We show that, under suitable restricted strong convexity and smoothness assumptions for the loss, the resulting estimator achieves the squared-error risk up to a multiplicative constant that is independent of . In contrast, previous polynomial-time algorithms have only been shown to achieve this guarantee in more specialized settings, or under additional assumptions for and/or the sparsity pattern of . As applications of our general framework, we apply our results to the examples of linear models and generalized linear models with random design.
View on arXiv