150

Neural networks and rational functions

International Conference on Machine Learning (ICML), 2017
Abstract

Neural networks and rational functions efficiently approximate each other. In more detail, it is shown here that for any ReLU network, there exists a rational function of degree O(polylog(1/ϵ))O(\text{polylog}(1/\epsilon)) which is ϵ\epsilon-close, and similarly for any rational function there exists a ReLU network of size O(polylog(1/ϵ))O(\text{polylog}(1/\epsilon)) which is ϵ\epsilon-close. By contrast, polynomials need degree Ω(poly(1/ϵ))\Omega(\text{poly}(1/\epsilon)) to approximate even a single ReLU. When converting a ReLU network to a rational function as above, the hidden constants depend exponentially on the number of layers, which is shown to be tight; in other words, a compositional representation can be beneficial even for rational functions.

View on arXiv
Comments on this paper