We analyze the optimization landscape of a recently introduced tunable class of loss functions called -loss, , in the logistic model. This family encapsulates the exponential loss (), the log-loss (), and the 0-1 loss () and contains compelling properties that enable the practitioner to discern among a host of operating conditions relevant to emerging learning methods. Specifically, we study the evolution of the optimization landscape of -loss with respect to using tools drawn from the study of strictly-locally-quasi-convex functions in addition to geometric techniques. We interpret these results in terms of optimization complexity via normalized gradient descent.
View on arXiv