254

Nonclosedness of the Set of Neural Networks in Sobolev Space

Abstract

We examine the closedness of the set of realized neural networks of a fixed architecture in Sobolev space. For an exactly mm-times differentiable activation function ρ\rho, we construct a sequence of neural networks (Φn)nN(\Phi_n)_{n \in \mathbb{N}} whose realizations converge in order-(m1)(m-1) Sobolev norm to a function that cannot be realized exactly by a neural network. Thus, the set of realized neural networks is not closed in the order-(m1)(m-1) Sobolev space Wm1,pW^{m-1,p}. We further show that this set is not closed in Wm,pW^{m,p} under slightly stronger conditions on the mm-th derivative of ρ\rho. For a real analytic activation function, we show that the set of realized neural networks is not closed in Wk,pW^{k,p} for any kNk \in \mathbb{N}. These results suggest that training a network to approximate a target function in Sobolev norm does not prevent parameter explosion. Finally, we present experimental results demonstrating that parameter explosion occurs in stochastic training regardless of the norm under which the network is trained. However, the network is still capable of closely approximating a non-network target function with network parameters that grow at a manageable rate.

View on arXiv
Comments on this paper