217

Unraveling the Black-box Magic: An Analysis of Neural Networks' Dynamic Local Extrema

Main:17 Pages
8 Figures
Bibliography:2 Pages
1 Tables
Abstract

We point out that neural networks are not black boxes, and their generalization stems from the ability to dynamically map a dataset to the local extrema of the model function. We further prove that the number of local extrema in a neural network is positively correlated with the number of its parameters, and on this basis, we give a new algorithm that is different from the back-propagation algorithm, which we call the extremum-increment algorithm. Some difficult situations, such as gradient vanishing and overfitting, can be reasonably explained and dealt with in this framework.

View on arXiv
Comments on this paper