441

Learned-norm pooling for deep neural networks

Abstract

In this paper we proposed a novel nonlinear unit, which is called as LpL_p unit, for a multi-layer perceptron (MLP). The proposed LpL_p unit receives signal from several projections of the layer below and computes the normalized LpL_p norm. We notice two interesting interpretations of the LpL_p unit. First, we note that the proposed unit is a generalization of a number of conventional pooling operators such as average, root-mean-square and max pooling widely used in, for instance, convolutional neural networks(CNN), HMAX models and neocognitrons. Furthermore, under certain constraints, the LpL_p unit is a generalization of the recently proposed maxout unit (Goodfellow et al, 2013) which achieved the state-of-the-art object recognition results on a number of benchmark datasets. Second, we provide a geometrical interpretation of the activation function. Each LpL_p unit defines a spherical boundary, with its exact shape defined by the order pp. We claim that this makes it possible to obtain arbitrarily shaped, curved boundaries more efficiently by combining just a few LpL_p units of different orders. We empirically evaluate the proposed LpL_p units on a number of datasets and show that MLPs consisting of the LpL_p units achieves the state-of-the-art results on a number of benchmark datasets.

View on arXiv
Comments on this paper