How to Shrink Confidence Sets for Many Equivalent Discrete Distributions?

We consider the situation when a learner faces a set of unknown discrete distributions defined over a common alphabet , and can build for each distribution an individual high-probability confidence set thanks to observations sampled from . The set is structured: each distribution is obtained from the same common, but unknown, distribution q via applying an unknown permutation to . We call this \emph{permutation-equivalence}. The goal is to build refined confidence sets \emph{exploiting} this structural property. Like other popular notions of structure (Lipschitz smoothness, Linearity, etc.) permutation-equivalence naturally appears in machine learning problems, and to benefit from its potential gain calls for a specific approach. We present a strategy to effectively exploit permutation-equivalence, and provide a finite-time high-probability bound on the size of the refined confidence sets output by the strategy. Since a refinement is not possible for too few observations in general, under mild technical assumptions, our finite-time analysis establish when the number of observations are large enough so that the output confidence sets improve over initial individual sets. We carefully characterize this event and the corresponding improvement. Further, our result implies that the size of confidence sets shrink at asymptotic rates of and , respectively for elements inside and outside the support of q, when the size of each individual confidence set shrinks at respective rates of and . We illustrate the practical benefit of exploiting permutation equivalence on a reinforcement learning task.
View on arXiv