22
2

On the optimal approximation of Sobolev and Besov functions using deep ReLU neural networks

Yunfei Yang
Abstract

This paper studies the problem of how efficiently functions in the Sobolev spaces Ws,q([0,1]d)\mathcal{W}^{s,q}([0,1]^d) and Besov spaces Bq,rs([0,1]d)\mathcal{B}^s_{q,r}([0,1]^d) can be approximated by deep ReLU neural networks with width WW and depth LL, when the error is measured in the Lp([0,1]d)L^p([0,1]^d) norm. This problem has been studied by several recent works, which obtained the approximation rate O((WL)2s/d)\mathcal{O}((WL)^{-2s/d}) up to logarithmic factors when p=q=p=q=\infty, and the rate O(L2s/d)\mathcal{O}(L^{-2s/d}) for networks with fixed width when the Sobolev embedding condition 1/q1/p<s/d1/q -1/p<s/d holds. We generalize these results by showing that the rate O((WL)2s/d)\mathcal{O}((WL)^{-2s/d}) indeed holds under the Sobolev embedding condition. It is known that this rate is optimal up to logarithmic factors. The key tool in our proof is a novel encoding of sparse vectors by using deep ReLU neural networks with varied width and depth, which may be of independent interest.

View on arXiv
Comments on this paper