ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2408.08494
25
2

Optimal Sketching for Residual Error Estimation for Matrix and Vector Norms

16 August 2024
Yi Li
Honghao Lin
David P. Woodruff
ArXiv (abs)PDFHTML
Abstract

We study the problem of residual error estimation for matrix and vector norms using a linear sketch. Such estimates can be used, for example, to quickly assess how useful a more expensive low-rank approximation computation will be. The matrix case concerns the Frobenius norm and the task is to approximate the kkk-residual ∥A−Ak∥F\|A - A_k\|_F∥A−Ak​∥F​ of the input matrix AAA within a (1+ϵ)(1+\epsilon)(1+ϵ)-factor, where AkA_kAk​ is the optimal rank-kkk approximation. We provide a tight bound of Θ(k2/ϵ4)\Theta(k^2/\epsilon^4)Θ(k2/ϵ4) on the size of bilinear sketches, which have the form of a matrix product SATSATSAT. This improves the previous O(k2/ϵ6)O(k^2/\epsilon^6)O(k2/ϵ6) upper bound in (Andoni et al. SODA 2013) and gives the first non-trivial lower bound, to the best of our knowledge. In our algorithm, our sketching matrices SSS and TTT can both be sparse matrices, allowing for a very fast update time. We demonstrate that this gives a substantial advantage empirically, for roughly the same sketch size and accuracy as in previous work. For the vector case, we consider the ℓp\ell_pℓp​-norm for p>2p>2p>2, where the task is to approximate the kkk-residual ∥x−xk∥p\|x - x_k\|_p∥x−xk​∥p​ up to a constant factor, where xkx_kxk​ is the optimal kkk-sparse approximation to xxx. Such vector norms are frequently studied in the data stream literature and are useful for finding frequent items or so-called heavy hitters. We establish an upper bound of O(k2/pn1−2/ppoly⁡(log⁡n))O(k^{2/p}n^{1-2/p}\operatorname{poly}(\log n))O(k2/pn1−2/ppoly(logn)) for constant ϵ\epsilonϵ on the dimension of a linear sketch for this problem. Our algorithm can be extended to the ℓp\ell_pℓp​ sparse recovery problem with the same sketching dimension, which seems to be the first such bound for p>2p > 2p>2. We also show an Ω(k2/pn1−2/p)\Omega(k^{2/p}n^{1-2/p})Ω(k2/pn1−2/p) lower bound for the sparse recovery problem, which is tight up to a poly(log⁡n)\mathrm{poly}(\log n)poly(logn) factor.

View on arXiv
Comments on this paper