398
v1v2 (latest)

Optimal approximation of continuous functions by very deep ReLU networks

Abstract

We consider approximations of general continuous functions on finite-dimensional cubes by general deep ReLU neural networks and study the approximation rates with respect to the modulus of continuity of the function and the total number of weights WW in the network. We establish the complete phase diagram of feasible approximation rates and show that it includes two distinct phases. One phase corresponds to slower approximations that can be achieved with constant-depth networks and continuous weight assignments. The other phase provides faster approximations at the cost of depths necessarily growing as a power law LWα,0<α1,L\sim W^{\alpha}, 0<\alpha\le 1, and with necessarily discontinuous weight assignments. In particular, we prove that constant-width fully-connected networks of depth LWL\sim W provide the fastest possible approximation rate ff~=O(ωf(O(W2/ν)))\|f-\widetilde f\|_\infty = O(\omega_f(O(W^{-2/\nu}))) that cannot be achieved with less deep networks.

View on arXiv
Comments on this paper