332

Supermodular f-divergences and bounds on lossy compression and generalization error with mutual f-information

IEEE Transactions on Information Theory (IEEE Trans. Inf. Theory), 2022
Abstract

In this paper, we introduce super-modular \mf\mf-divergences and provide three applications for them: (i) we introduce Sanov's upper bound on the tail probability of sum of independent random variables based on super-modular \mf\mf-divergence and show that our generalized Sanov's bound strictly improves over ordinary one, (ii) we consider the lossy compression problem which studies the set of achievable rates for a given distortion and code length. We extend the rate-distortion function using mutual \mf\mf-information and provide new and strictly better bounds on achievable rates in the finite blocklength regime using super-modular \mf\mf-divergences, and (iii) we provide a connection between the generalization error of algorithms with bounded input/output mutual \mf\mf-information and a generalized rate-distortion problem. This connection allows us to bound the generalization error of learning algorithms using lower bounds on the rate-distortion function. Our bound is based on a new lower bound on the rate-distortion function that (for some examples) strictly improves over previously best-known bounds. Moreover, super-modular \mf\mf-divergences are utilized to reduce the dimension of the problem and obtain single-letter bounds.

View on arXiv
Comments on this paper