257

Universal approximations of permutation invariant/equivariant functions by deep neural networks

Abstract

In this paper,we develop a theory of the relationship between permutation (SnS_n-) invariant/equivariant functions and deep neural networks. As a result, we prove an permutation invariant/equivariant version of the universal approximation theorem, i.e SnS_n-invariant/equivariant deep neural networks. The equivariant models are consist of stacking standard single-layer neural networks Zi:XYZ_i:X \to Y for which every ZiZ_i is SnS_n-equivariant with respect to the actions of SnS_n . The invariant models are consist of stacking equivariant models and standard single-layer neural networks Zi:XYZ_i:X \to Y for which every ZiZ_i is SnS_n-invariant with respect to the actions of SnS_n . These are universal approximators to SnS_n-invariant/equivariant functions. The above notation is mathematically natural generalization of the models in \cite{deepsets}. We also calculate the number of free parameters appeared in these models. As a result, the number of free parameters appeared in these models is much smaller than the one of the usual models. Hence, we conclude that although the free parameters of the invariant/equivarint models are exponentially fewer than the one of the usual models, the invariant/equivariant models can approximate the invariant/equivariant functions to arbitrary accuracy. This gives us an understanding of why the invariant/equivariant models designed in [Zaheer et al. 2018] work well.

View on arXiv
Comments on this paper