53
36

Sample-Optimal Low-Rank Approximation of Distance Matrices

Abstract

A distance matrix ARn×mA \in \mathbb R^{n \times m} represents all pairwise distances, Aij=d(xi,yj)A_{ij}=\mathrm{d}(x_i,y_j), between two point sets x1,...,xnx_1,...,x_n and y1,...,ymy_1,...,y_m in an arbitrary metric space (Z,d)(\mathcal Z, \mathrm{d}). Such matrices arise in various computational contexts such as learning image manifolds, handwriting recognition, and multi-dimensional unfolding. In this work we study algorithms for low-rank approximation of distance matrices. Recent work by Bakshi and Woodruff (NeurIPS 2018) showed it is possible to compute a rank-kk approximation of a distance matrix in time O((n+m)1+γ)poly(k,1/ϵ)O((n+m)^{1+\gamma}) \cdot \mathrm{poly}(k,1/\epsilon), where ϵ>0\epsilon>0 is an error parameter and γ>0\gamma>0 is an arbitrarily small constant. Notably, their bound is sublinear in the matrix size, which is unachievable for general matrices. We present an algorithm that is both simpler and more efficient. It reads only O((n+m)k/ϵ)O((n+m) k/\epsilon) entries of the input matrix, and has a running time of O(n+m)poly(k,1/ϵ)O(n+m) \cdot \mathrm{poly}(k,1/\epsilon). We complement the sample complexity of our algorithm with a matching lower bound on the number of entries that must be read by any algorithm. We provide experimental results to validate the approximation quality and running time of our algorithm.

View on arXiv
Comments on this paper