76

On Weak-to-Strong Generalization and f-Divergence

Main:9 Pages
2 Figures
Bibliography:4 Pages
2 Tables
Appendix:6 Pages
Abstract

Weak-to-strong generalization (W2SG) has emerged as a promising paradigm for stimulating the capabilities of strong pre-trained models by leveraging supervision from weaker supervisors. To improve the performance of the strong model, existing methods often require additional weak models or complex procedures, leading to substantial computational and memory overhead. Motivated by the effectiveness of ff-divergence loss in various machine learning domains, we introduce ff-divergence as an information-theoretic loss function framework in W2SG. Our theoretical analysis reveals fundamental limitations and equivalence of different ff-divergence losses in W2SG, supported by sample complexity bounds and information-theoretic insights. We empirically demonstrate that ff-divergence loss, which generalizes widely-used metrics like KL divergence, effectively improves generalization and noise tolerance of the strong model in practice.

View on arXiv
Comments on this paper