23
42

On the emergence of simplex symmetry in the final and penultimate layers of neural network classifiers

Abstract

A recent numerical study observed that neural network classifiers enjoy a large degree of symmetry in the penultimate layer. Namely, if h(x)=Af(x)+bh(x) = Af(x) +b where AA is a linear map and ff is the output of the penultimate layer of the network (after activation), then all data points xi,1,,xi,Nix_{i, 1}, \dots, x_{i, N_i} in a class CiC_i are mapped to a single point yiy_i by ff and the points yiy_i are located at the vertices of a regular k1k-1-dimensional standard simplex in a high-dimensional Euclidean space. We explain this observation analytically in toy models for highly expressive deep neural networks. In complementary examples, we demonstrate rigorously that even the final output of the classifier hh is not uniform over data samples from a class CiC_i if hh is a shallow network (or if the deeper layers do not bring the data samples into a convenient geometric configuration).

View on arXiv
Comments on this paper