64
203

Information-theoretic limits of selecting binary graphical models in high dimensions

Abstract

The problem of graphical model selection is to correctly estimate the graph structure of a Markov random field given samples from the underlying distribution. We analyze the information-theoretic limitations of the problem of graph selection for binary Markov random fields under high-dimensional scaling, in which the graph size pp and the number of edges kk, and/or the maximal node degree dd are allowed to increase to infinity as a function of the sample size nn. For pairwise binary Markov random fields, we derive both necessary and sufficient conditions for correct graph selection over the class Gp,k\mathcal{G}_{p,k} of graphs on pp vertices with at most kk edges, and over the class Gp,d\mathcal{G}_{p,d} of graphs on pp vertices with maximum degree at most dd. For the class Gp,k\mathcal{G}_{p, k}, we establish the existence of constants cc and cc' such that if \numobs<cklogp\numobs < c k \log p, any method has error probability at least 1/2 uniformly over the family, and we demonstrate a graph decoder that succeeds with high probability uniformly over the family for sample sizes \numobs>ck2logp\numobs > c' k^2 \log p. Similarly, for the class Gp,d\mathcal{G}_{p,d}, we exhibit constants cc and cc' such that for n<cd2logpn < c d^2 \log p, any method fails with probability at least 1/2, and we demonstrate a graph decoder that succeeds with high probability for n>cd3logpn > c' d^3 \log p.

View on arXiv
Comments on this paper