201

Rivalry of Two Families of Algorithms for Memory-Restricted Streaming PCA

International Conference on Artificial Intelligence and Statistics (AISTATS), 2015
Abstract

We study the problem of recovering the subspace spanned by the first kk principal components of dd-dimensional data under the streaming setting, with a memory bound of O(kd)\mathcal{O}(kd). Two families of algorithms are known for this problem. The first family is based on stochastic gradient descent, but no convergence proof of its existing algorithms was previously known when k>1k>1. The second family is based on the power method over blocks of data, but setting the block size for its existing algorithms is not an easy task. In this paper, we analyze the convergence rate of a representative algorithm in the first family~\cite{oja} for the general k>1k>1 case. Moreover, we propose a novel algorithm for the second family that sets the block sizes automatically and dynamically with faster convergence rate. We then conduct empirical studies that fairly compare the two families on real-world data. The studies reveal the advantages and disadvantages of these two families.

View on arXiv
Comments on this paper