Beyond Moments: Robustly Learning Affine Transformations with Asymptotically Optimal Error

We present a polynomial-time algorithm for robustly learning an unknown affine transformation of the standard hypercube from samples, an important and well-studied setting for independent component analysis (ICA). Specifically, given an -corrupted sample from a distribution obtained by applying an unknown affine transformation to the uniform distribution on a -dimensional hypercube , our algorithm constructs such that the total variation distance of the distribution from is using poly time and samples. Total variation distance is the information-theoretically strongest possible notion of distance in our setting and our recovery guarantees in this distance are optimal up to the absolute constant factor multiplying . In particular, if the columns of are normalized to be unit length, our total variation distance guarantee implies a bound on the sum of the distances between the column vectors of and , . In contrast, the strongest known prior results only yield a (relative) bound on the distance between individual 's and their estimates and translate into an bound on the total variation distance. Our key innovation is a new approach to ICA (even to outlier-free ICA) that circumvents the difficulties in the classical method of moments and instead relies on a new geometric certificate of correctness of an affine transformation. Our algorithm is based on a new method that iteratively improves an estimate of the unknown affine transformation whenever the requirements of the certificate are not met.
View on arXiv