100
v1v2 (latest)

Drawback of Enforcing Equivariance and its Compensation via the Lens of Expressive Power

Yuzhu Chen
Tian Qin
Xinmei Tian
Fengxiang He
Dacheng Tao
Main:10 Pages
3 Figures
Bibliography:3 Pages
Appendix:6 Pages
Abstract

Equivariant neural networks encode symmetry as an inductive bias and have achieved strong empirical performance in wide domains. However, their expressive power remains not well understood. Focusing on 2-layer ReLU networks, this paper investigates the impact of equivariance constraints on the expressivity of equivariant and layer-wise equivariant networks. By examining the boundary hyperplanes and the channel vectors of ReLU networks, we construct an example showing that equivariance constraints could strictly limit expressive power. However, we demonstrate that this drawback can be compensated via enlarging the model size. Furthermore, we show that despite a larger model size, the resulting architecture could still correspond to a hypothesis space with lower complexity, implying superior generalizability for equivariant networks.

View on arXiv
Comments on this paper