54
2

Sharpened Error Bound for Random Sampling Based 2\ell_2 Regression

Abstract

Given a data matrix XRn×dX \in R^{n\times d} and a response vector yRny \in R^{n}, suppose n>dn>d, it costs O(nd2)O(n d^2) time and O(nd)O(n d) space to solve the least squares regression (LSR) problem. When nn and dd are both large, exactly solving the LSR problem is very expensive. When ndn \gg d, one feasible approach to accelerating LSR is to randomly embed yy and all columns of XX into the subspace RcR^c where cnc\ll n; the induced LSR problem has the same number of columns but much fewer number of rows, and the induced problem can be solved in O(cd2)O(c d^2) time and O(cd)O(c d) space. The leverage scores based sampling is an effective subspace embedding method and can be applied to accelerate LSR. It was shown previously that c=O(dϵ2logd)c = O(d \epsilon^{-2} \log d) is sufficient for achieving 1+ϵ1+\epsilon accuracy. In this paper we sharpen this error bound, showing that c=O(dlogd+dϵ1)c = O(d \log d + d \epsilon^{-1}) is enough for 1+ϵ1+\epsilon accuracy.

View on arXiv
Comments on this paper