Nearly-tight VC-dimension bounds for piecewise linear neural networks
Abstract
We prove new upper and lower bounds on the VC-dimension of deep neural networks with the ReLU activation function. These bounds are tight for almost the entire range of parameters. Letting be the number of weights and be the number of layers, we prove that the VC-dimension is and . This improves both the previously known upper bounds and lower bounds. In terms of the number of non-linear units, we prove a tight bound on the VC-dimension. All of these results generalize to arbitrary piecewise linear activation functions.
View on arXivComments on this paper
