55
v1v2 (latest)

Optimal Oblivious Subspace Embeddings with Near-optimal Sparsity

International Colloquium on Automata, Languages and Programming (ICALP), 2024
Main:68 Pages
3 Figures
Bibliography:3 Pages
Abstract

An oblivious subspace embedding is a random m×nm\times n matrix Π\Pi such that, for any dd-dimensional subspace, with high probability Π\Pi preserves the norms of all vectors in that subspace within a 1±ϵ1\pm\epsilon factor. In this work, we give an oblivious subspace embedding with the optimal dimension m=Θ(d/ϵ2)m=\Theta(d/\epsilon^2) that has a near-optimal sparsity of O~(1/ϵ)\tilde O(1/\epsilon) non-zero entries per column of Π\Pi. This is the first result to nearly match the conjecture of Nelson and Nguyen [FOCS 2013] in terms of the best sparsity attainable by an optimal oblivious subspace embedding, improving on a prior bound of O~(1/ϵ6)\tilde O(1/\epsilon^6) non-zeros per column [Chenakkod et al., STOC 2024]. We further extend our approach to the non-oblivious setting, proposing a new family of Leverage Score Sparsified embeddings with Independent Columns, which yield faster runtimes for matrix approximation and regression tasks.In our analysis, we develop a new method which uses a decoupling argument together with the cumulant method for bounding the edge universality error of isotropic random matrices. To achieve near-optimal sparsity, we combine this general-purpose approach with new traces inequalities that leverage the specific structure of our subspace embedding construction.

View on arXiv
Comments on this paper