8
3160

Deep Learning using Rectified Linear Units (ReLU)

Abstract

We introduce the use of rectified linear units (ReLU) as the classification function in a deep neural network (DNN). Conventionally, ReLU is used as an activation function in DNNs, with Softmax function as their classification function. However, there have been several studies on using a classification function other than Softmax, and this study is an addition to those. We accomplish this by taking the activation of the penultimate layer hn1h_{n - 1} in a neural network, then multiply it by weight parameters θ\theta to get the raw scores oio_{i}. Afterwards, we threshold the raw scores oio_{i} by 00, i.e. f(o)=max(0,oi)f(o) = \max(0, o_{i}), where f(o)f(o) is the ReLU function. We provide class predictions y^\hat{y} through argmax function, i.e. argmax f(x)f(x).

View on arXiv
Comments on this paper