391

Information-theoretic limits of selecting binary graphical models in high dimensions

IEEE Transactions on Information Theory (IEEE Trans. Inf. Theory), 2009
Abstract

The problem of graphical model selection is to correctly estimate the graph structure of a Markov random field given samples from the underlying distribution. We analyze the information-theoretic limitations of the problem of graph selection for binary Markov random fields under high-dimensional scaling, in which the graph size pp and the number of edges kk, and/or the maximal node degree dd are allowed to increase to infinity as a function of the sample size nn. For pairwise binary Markov random fields, we derive both necessary and sufficient conditions for correct graph selection over the class Gp,k\mathcal{G}_{p,k} of graphs on pp vertices with at most kk edges, and over the class Gp,d\mathcal{G}_{p,d} of graphs on pp vertices with maximum degree at most dd. For the class Gp,k\mathcal{G}_{p, k}, we establish the existence of constants cc and cc' such that if \numobs<cklogp\numobs < c k \log p, any method has error probability at least 1/2 uniformly over the family, and we demonstrate a graph decoder that succeeds with high probability uniformly over the family for sample sizes \numobs>ck2logp\numobs > c' k^2 \log p. Similarly, for the class Gp,d\mathcal{G}_{p,d}, we exhibit constants cc and cc' such that for n<cd2logpn < c d^2 \log p, any method fails with probability at least 1/2, and we demonstrate a graph decoder that succeeds with high probability for n>cd3logpn > c' d^3 \log p.

View on arXiv
Comments on this paper