153

Optimal Sketching for Residual Error Estimation for Matrix and Vector Norms

International Conference on Learning Representations (ICLR), 2024
Main:11 Pages
Bibliography:3 Pages
1 Tables
Abstract

We study the problem of residual error estimation for matrix and vector norms using a linear sketch. Such estimates can be used, for example, to quickly assess how useful a more expensive low-rank approximation computation will be. The matrix case concerns the Frobenius norm and the task is to approximate the kk-residual AAkF\|A - A_k\|_F of the input matrix AA within a (1+ϵ)(1+\epsilon)-factor, where AkA_k is the optimal rank-kk approximation. We provide a tight bound of Θ(k2/ϵ4)\Theta(k^2/\epsilon^4) on the size of bilinear sketches, which have the form of a matrix product SATSAT. This improves the previous O(k2/ϵ6)O(k^2/\epsilon^6) upper bound in (Andoni et al. SODA 2013) and gives the first non-trivial lower bound, to the best of our knowledge. In our algorithm, our sketching matrices SS and TT can both be sparse matrices, allowing for a very fast update time. We demonstrate that this gives a substantial advantage empirically, for roughly the same sketch size and accuracy as in previous work. For the vector case, we consider the p\ell_p-norm for p>2p>2, where the task is to approximate the kk-residual xxkp\|x - x_k\|_p up to a constant factor, where xkx_k is the optimal kk-sparse approximation to xx. Such vector norms are frequently studied in the data stream literature and are useful for finding frequent items or so-called heavy hitters. We establish an upper bound of O(k2/pn12/ppoly(logn))O(k^{2/p}n^{1-2/p}\operatorname{poly}(\log n)) for constant ϵ\epsilon on the dimension of a linear sketch for this problem. Our algorithm can be extended to the p\ell_p sparse recovery problem with the same sketching dimension, which seems to be the first such bound for p>2p > 2. We also show an Ω(k2/pn12/p)\Omega(k^{2/p}n^{1-2/p}) lower bound for the sparse recovery problem, which is tight up to a poly(logn)\mathrm{poly}(\log n) factor.

View on arXiv
Comments on this paper