199
v1v2 (latest)

Constructive Universal Approximation and Finite Sample Memorization by Narrow Deep ReLU Networks

26 Figures
Appendix:46 Pages
Abstract

We present a fully constructive analysis of deep ReLU neural networks for classification and function approximation tasks. First, we prove that any dataset with NN distinct points in Rd\mathbb{R}^d and MM output classes can be exactly classified using a multilayer perceptron (MLP) of width 22 and depth at most 2N+4M12N + 4M - 1, with all network parameters constructed explicitly. This result is sharp with respect to width and is interpreted through the lens of simultaneous or ensemble controllability in discrete nonlinear dynamics.Second, we show that these explicit constructions yield uniform bounds on the parameter norms and, in particular, provide upper estimates for minimizers of standard regularized training loss functionals in supervised learning. As the regularization parameter vanishes, the trained networks converge to exact classifiers with bounded norm, explaining the effectiveness of overparameterized training in the small-regularization regime.We also prove a universal approximation theorem in Lp(Ω;R+)L^p(\Omega; \mathbb{R}_+) for any bounded domain ΩRd\Omega \subset \mathbb{R}^d and p[1,)p \in [1, \infty), using MLPs of fixed width d+1d + 1. The proof is constructive, geometrically motivated, and provides explicit estimates on the network depth when the target function belongs to the Sobolev space W1,pW^{1,p}. We also extend the approximation and depth estimation results to Lp(Ω;Rm)L^p(\Omega; \mathbb{R}^m) for any m1m \geq 1.Our results offer a unified and interpretable framework connecting controllability, expressivity, and training dynamics in deep neural networks.

View on arXiv
Comments on this paper