215

Compressed Dictionary Learning

Abstract

In this paper we show that the computational complexity of the Iterative Thresholding and K-Residual-Means (ITKrM) algorithm for dictionary learning can be significantly reduced by using dimensionality reduction techniques based on the Johnson-Lindenstrauss Lemma. We introduce the Iterative Compressed-Thresholding and K-Means (IcTKM) algorithm for fast dictionary learning and study its convergence properties. We show that IcTKM can locally recover a generating dictionary with low computational complexity up to a target error ε~\tilde{\varepsilon} by compressing dd-dimensional training data into m<dm < d dimensions, where mm is proportional to logd\log d and inversely proportional to the distortion level δ\delta incurred by compressing the data. Increasing the distortion level δ\delta reduces the computational complexity of IcTKM at the cost of an increased recovery error and reduced admissible sparsity level for the training data. For generating dictionaries comprised of KK atoms, we show that IcTKM can stably recover the dictionary with distortion levels up to the order δO(1/logK)\delta \leq O(1/\sqrt{\log K}). The compression effectively shatters the data dimension bottleneck in the computational cost of the ITKrM algorithm. For training data with sparsity levels SO(K2/3)S \leq O(K^{2/3}), ITKrM can locally recover the dictionary with a computational cost that scales as O(dKlog(ε~1))O(d K \log(\tilde{\varepsilon}^{-1})) per training signal. We show that for these same sparsity levels the computational cost can be brought down to O(log5(d)Klog(ε~1))O(\log^5 (d) K \log(\tilde{\varepsilon}^{-1})) with IcTKM, a significant reduction when high-dimensional data is considered. Our theoretical results are complemented with numerical simulations which demonstrate that IcTKM is a powerful, low-cost algorithm for learning dictionaries from high-dimensional data sets.

View on arXiv
Comments on this paper