Sharpened Error Bound for Random Sampling Based Regression

Abstract
Given a data matrix and a response vector , suppose , it costs time and space to solve the least squares regression (LSR) problem. When and are both large, exactly solving the LSR problem is very expensive. When , one feasible approach to accelerating LSR is to randomly embed and all columns of into the subspace where ; the induced LSR problem has the same number of columns but much fewer number of rows, and the induced problem can be solved in time and space. The leverage scores based sampling is an effective subspace embedding method and can be applied to accelerate LSR. It was shown previously that is sufficient for achieving accuracy. In this paper we sharpen this error bound, showing that is enough for accuracy.
View on arXivComments on this paper