ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1512.09080
167
120
v1v2v3v4 (latest)

Detection in the stochastic block model with multiple clusters: proof of the achievability conjectures, acyclic BP, and the information-computation gap

30 December 2015
Emmanuel Abbe
Colin Sandon
ArXiv (abs)PDFHTML
Abstract

In a paper that initiated the modern study of the stochastic block model, Decelle, Krzakala, Moore and Zdeborov\'a made a fascinating conjecture: Denote by kkk the number of balanced communities, a/na/na/n the probability of connecting inside clusters and b/nb/nb/n across clusters, and set SNR=(a−b)2/(k(a+(k−1)b))\mathrm{SNR}=(a-b)^2/(k(a+(k-1)b))SNR=(a−b)2/(k(a+(k−1)b)); for any k≥2k \geq 2k≥2, it is possible to detect efficiently communities whenever SNR>1\mathrm{SNR}>1SNR>1 (the KS bound), whereas for k≥5k\geq 5k≥5, it is possible to detect communities information-theoretically for some SNR<1\mathrm{SNR}<1SNR<1. Massouli\'e, Mossel et al.\ and Bordenave et al.\ succeeded in proving that the KS bound is efficiently achievable for k=2k=2k=2, while Mossel et al.\ proved that the KS bound cannot be crossed information-theoretically for k=2k=2k=2. The above conjecture remained open beyond two communities. This paper proves this conjecture. Specifically, an acyclic belief propagation (ABP) algorithm is developed and proved to detect communities for any kkk whenever SNR>1\mathrm{SNR}>1SNR>1, i.e., down the KS bound, with a complexity of O(nlog⁡n)O(n \log n)O(nlogn). The parameters are also learned efficiently down the KS bound. Thus ABP improves upon the state-of-the-art both in terms of complexity and universality in achieving the KS bound. This also gives a rigorous instance where message passing succeeds with optimal guarantees while handling cycles and a random initialization. Further, a non-efficient algorithm sampling a typical clustering is shown to break down the KS bound at k=5k=5k=5. The gap between the KS bound and the information-theoretic bound is shown to be large in some cases; if a=0a=0a=0, the KS bound reads b≳k2b \gtrsim k^2b≳k2 whereas the information-theoretic bound reads b≳kln⁡(k)b \gtrsim k \ln(k)b≳kln(k). The results extend to general SBMs.

View on arXiv
Comments on this paper