New Statistical and Computational Results for Learning Junta Distributions

We study the problem of learning junta distributions on , where a distribution is a -junta if its probability mass function depends on a subset of at most variables. We make two main contributions:- We show that learning -junta distributions is \emph{computationally} equivalent to learning -parity functions with noise (LPN), a landmark problem in computational learning theory.- We design an algorithm for learning junta distributions whose statistical complexity is optimal, up to polylogarithmic factors. Computationally, our algorithm matches the complexity of previous (non-sample-optimal) algorithms.Combined, our two contributions imply that our algorithm cannot be significantly improved, statistically or computationally, barring a breakthrough for LPN.
View on arXiv@article{beretta2025_2505.05819, title={ New Statistical and Computational Results for Learning Junta Distributions }, author={ Lorenzo Beretta }, journal={arXiv preprint arXiv:2505.05819}, year={ 2025 } }