21
1

In-depth Analysis of Low-rank Matrix Factorisation in a Federated Setting

Abstract

We analyze a distributed algorithm to compute a low-rank matrix factorization on NN clients, each holding a local dataset SiRni×d\mathbf{S}^i \in \mathbb{R}^{n_i \times d}, mathematically, we seek to solve minUiRni×r,VRd×r12i=1NSiUiVF2min_{\mathbf{U}^i \in \mathbb{R}^{n_i\times r}, \mathbf{V}\in \mathbb{R}^{d \times r} } \frac{1}{2} \sum_{i=1}^N \|\mathbf{S}^i - \mathbf{U}^i \mathbf{V}^\top\|^2_{\text{F}}. Considering a power initialization of V\mathbf{V}, we rewrite the previous smooth non-convex problem into a smooth strongly-convex problem that we solve using a parallel Nesterov gradient descent potentially requiring a single step of communication at the initialization step. For any client ii in {1,,N}\{1, \dots, N\}, we obtain a global V\mathbf{V} in Rd×r\mathbb{R}^{d \times r} common to all clients and a local variable Ui\mathbf{U}^i in Rni×r\mathbb{R}^{n_i \times r}. We provide a linear rate of convergence of the excess loss which depends on σmax/σr\sigma_{\max} / \sigma_{r}, where σr\sigma_{r} is the rthr^{\mathrm{th}} singular value of the concatenation S\mathbf{S} of the matrices (Si)i=1N(\mathbf{S}^i)_{i=1}^N. This result improves the rates of convergence given in the literature, which depend on σmax2/σmin2\sigma_{\max}^2 / \sigma_{\min}^2. We provide an upper bound on the Frobenius-norm error of reconstruction under the power initialization strategy. We complete our analysis with experiments on both synthetic and real data.

View on arXiv
Comments on this paper