An -Poisson Multinomial Distribution (PMD) is the distribution of the sum of independent random vectors supported on the set of standard basis vectors in . We prove a structural characterization of these distributions, showing that, for all , any -Poisson multinomial random vector is -close, in total variation distance, to the sum of a discretized multidimensional Gaussian and an independent -Poisson multinomial random vector. Our structural characterization extends the multi-dimensional CLT of Valiant and Valiant, by simultaneously applying to all approximation requirements . In particular, it overcomes factors depending on and, importantly, the minimum eigenvalue of the PMD's covariance matrix from the distance to a multidimensional Gaussian random variable. We use our structural characterization to obtain an -cover, in total variation distance, of the set of all -PMDs, significantly improving the cover size of Daskalakis and Papadimitriou, and obtaining the same qualitative dependence of the cover size on and as the cover of Daskalakis and Papadimitriou. We further exploit this structure to show that -PMDs can be learned to within in total variation distance from samples, which is near-optimal in terms of dependence on and independent of . In particular, our result generalizes the single-dimensional result of Daskalakis, Diakonikolas, and Servedio for Poisson Binomials to arbitrary dimension.
View on arXiv