21
202

Complete Dictionary Recovery over the Sphere

Abstract

We consider the problem of recovering a complete (i.e., square and invertible) matrix A0\mathbf A_0, from YRn×p\mathbf Y \in \mathbb R^{n \times p} with Y=A0X0\mathbf Y = \mathbf A_0 \mathbf X_0, provided X0\mathbf X_0 is sufficiently sparse. This recovery problem is central to the theoretical understanding of dictionary learning, which seeks a sparse representation for a collection of input signals, and finds numerous applications in modern signal processing and machine learning. We give the first efficient algorithm that provably recovers A0\mathbf A_0 when X0\mathbf X_0 has O(n)O(n) nonzeros per column, under suitable probability model for X0\mathbf X_0. In contrast, prior results based on efficient algorithms provide recovery guarantees when X0\mathbf X_0 has only O(n1δ)O(n^{1-\delta}) nonzeros per column for any constant δ(0,1)\delta \in (0, 1). Our algorithmic pipeline centers around solving a certain nonconvex optimization problem with a spherical constraint, and hence is naturally phrased in the language of manifold optimization. To show this apparently hard problem is tractable, we first provide a geometric characterization of the high-dimensional objective landscape, which shows that with high probability there are no "spurious" local minima. This particular geometric structure allows us to design a Riemannian trust region algorithm over the sphere that provably converges to one local minimizer with an arbitrary initialization, despite the presence of saddle points. The geometric approach we develop here may also shed light on other problems arising from nonconvex recovery of structured signals.

View on arXiv
Comments on this paper