48
4

One-pass additive-error subset selection for p\ell_{p} subspace approximation

Abstract

We consider the problem of subset selection for p\ell_{p} subspace approximation, that is, to efficiently find a \emph{small} subset of data points such that solving the problem optimally for this subset gives a good approximation to solving the problem optimally for the original input. Previously known subset selection algorithms based on volume sampling and adaptive sampling \cite{DeshpandeV07}, for the general case of p[1,)p \in [1, \infty), require multiple passes over the data. In this paper, we give a one-pass subset selection with an additive approximation guarantee for p\ell_{p} subspace approximation, for any p[1,)p \in [1, \infty). Earlier subset selection algorithms that give a one-pass multiplicative (1+ϵ)(1+\epsilon) approximation work under the special cases. Cohen \textit{et al.} \cite{CohenMM17} gives a one-pass subset section that offers multiplicative (1+ϵ)(1+\epsilon) approximation guarantee for the special case of 2\ell_{2} subspace approximation. Mahabadi \textit{et al.} \cite{MahabadiRWZ20} gives a one-pass \emph{noisy} subset selection with (1+ϵ)(1+\epsilon) approximation guarantee for p\ell_{p} subspace approximation when p{1,2}p \in \{1, 2\}. Our subset selection algorithm gives a weaker, additive approximation guarantee, but it works for any p[1,)p \in [1, \infty).

View on arXiv
Comments on this paper