Distributed Consensus Algorithms in Sensor Networks: Link Failures and Channel Noise

The paper studies average consensus with random topologies (intermittent links) \emph{and} noisy channels. Consensus with noise in the network links leads to the bias-variance dilemma--running consensus for long reduces the bias of the final average estimate but increases its variance. We present two different compromises to this tradeoff: the algorithm modifies conventional consensus by forcing the weights to satisfy a \emph{persistence} condition (slowly decaying to zero); and the algorithm where the weights are constant but consensus is run for a fixed number of iterations , then it is restarted and rerun for a total of runs, and at the end averages the final states of the runs (Monte Carlo averaging). We use controlled Markov processes and stochastic approximation arguments to prove almost sure convergence of to the desired average (asymptotic unbiasedness) and compute explicitly the m.s.e. (variance) of the consensus limit. We show that represents the best of both worlds--low bias and low variance--at the cost of a slow convergence rate; rescaling the weights...
View on arXiv