120

ULU: A Unified Activation Function

Main:12 Pages
7 Figures
Bibliography:3 Pages
4 Tables
Abstract

We propose \textbf{ULU}, a novel non-monotonic, piecewise activation function defined as {f(x;α1),x<0;f(x;α2),x>=0}\{f(x;\alpha_1),x<0; f(x;\alpha_2),x>=0 \}, where f(x;α)=0.5x(tanh(αx)+1),α>0f(x;\alpha)=0.5x(tanh(\alpha x)+1),\alpha >0. ULU treats positive and negative inputs differently. Extensive experiments demonstrate ULU significantly outperforms ReLU and Mish across image classification and object detection tasks. Its variant Adaptive ULU (\textbf{AULU}) is expressed as {f(x;β12),x<0;f(x;β22),x>=0}\{f(x;\beta_1^2),x<0; f(x;\beta_2^2),x>=0 \}, where β1\beta_1 and β2\beta_2 are learnable parameters, enabling it to adapt its response separately for positive and negative inputs. Additionally, we introduce the LIB (Like Inductive Bias) metric from AULU to quantitatively measure the inductive bias of the model.

View on arXiv
Comments on this paper