ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2402.01238
11
1

Flexible Variational Information Bottleneck: Achieving Diverse Compression with a Single Training

2 February 2024
Sota Kudo
N. Ono
Shigehiko Kanaya
Ming Huang
ArXivPDFHTML
Abstract

Information Bottleneck (IB) is a widely used framework that enables the extraction of information related to a target random variable from a source random variable. In the objective function, IB controls the trade-off between data compression and predictiveness through the Lagrange multiplier β\betaβ. Traditionally, to find the trade-off to be learned, IB requires a search for β\betaβ through multiple training cycles, which is computationally expensive. In this study, we introduce Flexible Variational Information Bottleneck (FVIB), an innovative framework for classification task that can obtain optimal models for all values of β\betaβ with single, computationally efficient training. We theoretically demonstrate that across all values of reasonable β\betaβ, FVIB can simultaneously maximize an approximation of the objective function for Variational Information Bottleneck (VIB), the conventional IB method. Then we empirically show that FVIB can learn the VIB objective as effectively as VIB. Furthermore, in terms of calibration performance, FVIB outperforms other IB and calibration methods by enabling continuous optimization of β\betaβ. Our codes are available at https://github.com/sotakudo/fvib.

View on arXiv
Comments on this paper