52
v1v2 (latest)

Implicit Hypothesis Testing and Divergence Preservation in Neural Network Representations

Kadircan Aksoy
Protim Bhattacharjee
Peter Jung
Main:6 Pages
9 Figures
Bibliography:2 Pages
1 Tables
Appendix:5 Pages
Abstract

We study the supervised training dynamics of neural classifiers through the lens of binary hypothesis testing. We model classification as a set of binary tests between class-conditional distributions of representations and empirically show that, along training trajectories, well-generalizing networks increasingly align with Neyman-Pearson optimal decision rules via monotonic improvements in KL divergence that relate to error rate exponents. We finally discuss how this yields an explanation and possible training or regularization strategies for different classes of neural networks.

View on arXiv
Comments on this paper