We study the problem of residual error estimation for matrix and vector norms using a linear sketch. Such estimates can be used, for example, to quickly assess how useful a more expensive low-rank approximation computation will be. The matrix case concerns the Frobenius norm and the task is to approximate the -residual of the input matrix within a -factor, where is the optimal rank- approximation. We provide a tight bound of on the size of bilinear sketches, which have the form of a matrix product . This improves the previous upper bound in (Andoni et al. SODA 2013) and gives the first non-trivial lower bound, to the best of our knowledge. In our algorithm, our sketching matrices and can both be sparse matrices, allowing for a very fast update time. We demonstrate that this gives a substantial advantage empirically, for roughly the same sketch size and accuracy as in previous work. For the vector case, we consider the -norm for , where the task is to approximate the -residual up to a constant factor, where is the optimal -sparse approximation to . Such vector norms are frequently studied in the data stream literature and are useful for finding frequent items or so-called heavy hitters. We establish an upper bound of for constant on the dimension of a linear sketch for this problem. Our algorithm can be extended to the sparse recovery problem with the same sketching dimension, which seems to be the first such bound for . We also show an lower bound for the sparse recovery problem, which is tight up to a factor.
View on arXiv