15
6

Neural networks with superexpressive activations and integer weights

Abstract

An example of an activation function σ\sigma is given such that networks with activations {σ,}\{\sigma, \lfloor\cdot\rfloor\}, integer weights and a fixed architecture depending on dd approximate continuous functions on [0,1]d[0,1]^d. The range of integer weights required for ε\varepsilon-approximation of H\"older continuous functions is derived, which leads to a convergence rate of order n2β2β+dlog2nn^{\frac{-2\beta}{2\beta+d}}\log_2n for neural network regression estimation of unknown β\beta-H\"older continuous function with given nn samples.

View on arXiv
Comments on this paper