260
v1v2v3 (latest)

Approximation Rates for Shallow ReLUk^k Neural Networks on Sobolev Spaces via the Radon Transform

Main:10 Pages
Bibliography:3 Pages
1 Tables
Abstract

Let ΩRd\Omega\subset \mathbb{R}^d be a bounded domain. We consider the problem of how efficiently shallow neural networks with the ReLUk^k activation function can approximate functions from Sobolev spaces Ws(Lp(Ω))W^s(L_p(\Omega)) with error measured in the Lq(Ω)L_q(\Omega)-norm. Utilizing the Radon transform and recent results from discrepancy theory, we provide a simple proof of nearly optimal approximation rates in a variety of cases, including when qpq\leq p, p2p\geq 2, and sk+(d+1)/2s \leq k + (d+1)/2. The rates we derive are optimal up to logarithmic factors, and significantly generalize existing results. An interesting consequence is that the adaptivity of shallow ReLUk^k neural networks enables them to obtain optimal approximation rates for smoothness up to order s=k+(d+1)/2s = k + (d+1)/2, even though they represent piecewise polynomials of fixed degree kk.

View on arXiv
Comments on this paper