438
v1v2v3 (latest)

Multiplicative Logit Adjustment Approximates Neural-Collapse-Aware Decision Boundary Adjustment

International Conference on Learning Representations (ICLR), 2024
Main:10 Pages
7 Figures
Bibliography:5 Pages
14 Tables
Appendix:18 Pages
Abstract

Real-world data distributions are often highly skewed. This has spurred a growing body of research on long-tailed recognition, aimed at addressing the imbalance in training classification models. Among the methods studied, multiplicative logit adjustment (MLA) stands out as a simple and effective method. What theoretical foundation explains the effectiveness of this heuristic method? We provide a justification for the effectiveness of MLA with the following two-step process. First, we develop a theory that adjusts optimal decision boundaries by estimating feature spread on the basis of neural collapse. Second, we demonstrate that MLA approximates this optimal method. Additionally, through experiments on long-tailed datasets, we illustrate the practical usefulness of MLA under more realistic conditions. We also offer experimental insights to guide the tuning of MLA hyperparameters.

View on arXiv
Comments on this paper