13
0

Most Neural Networks Are Almost Learnable

Abstract

We present a PTAS for learning random constant-depth networks. We show that for any fixed ϵ>0\epsilon>0 and depth ii, there is a poly-time algorithm that for any distribution on dSd1\sqrt{d} \cdot \mathbb{S}^{d-1} learns random Xavier networks of depth ii, up to an additive error of ϵ\epsilon. The algorithm runs in time and sample complexity of (dˉ)poly(ϵ1)(\bar{d})^{\mathrm{poly}(\epsilon^{-1})}, where dˉ\bar d is the size of the network. For some cases of sigmoid and ReLU-like activations the bound can be improved to (dˉ)polylog(ϵ1)(\bar{d})^{\mathrm{polylog}(\epsilon^{-1})}, resulting in a quasi-poly-time algorithm for learning constant depth random networks.

View on arXiv
Comments on this paper