79
23
v1v2 (latest)

Communication Complexity of Estimating Correlations

Abstract

We characterize the communication complexity of the following distributed estimation problem. Alice and Bob observe infinitely many iid copies of ρ\rho-correlated unit-variance (Gaussian or ±1\pm1 binary) random variables, with unknown ρ[1,1]\rho\in[-1,1]. By interactively exchanging kk bits, Bob wants to produce an estimate ρ^\hat\rho of ρ\rho. We show that the best possible performance (optimized over interaction protocol Π\Pi and estimator ρ^\hat \rho) satisfies infΠρ^supρE[ρρ^2]=1k(12ln2+o(1))\inf_{\Pi \hat\rho}\sup_\rho \mathbb{E} [|\rho-\hat\rho|^2] = \tfrac{1}{k} (\frac{1}{2 \ln 2} + o(1)). Curiously, the number of samples in our achievability scheme is exponential in kk; by contrast, a naive scheme exchanging kk samples achieves the same Ω(1/k)\Omega(1/k) rate but with a suboptimal prefactor. Our protocol achieving optimal performance is one-way (non-interactive). We also prove the Ω(1/k)\Omega(1/k) bound even when ρ\rho is restricted to any small open sub-interval of [1,1][-1,1] (i.e. a local minimax lower bound). Our proof techniques rely on symmetric strong data-processing inequalities and various tensorization techniques from information-theoretic interactive common-randomness extraction. Our results also imply an Ω(n)\Omega(n) lower bound on the information complexity of the Gap-Hamming problem, for which we show a direct information-theoretic proof.

View on arXiv
Comments on this paper