70
28

Hoeffding's lemma for Markov Chains and its applications to statistical learning

Abstract

We establish the counterpart of Hoeffding's lemma for Markov dependent random variables. Specifically, if a stationary Markov chain {Xi}i1\{X_i\}_{i \ge 1} with invariant measure π\pi admits an L2(π)\mathcal{L}_2(\pi)-spectral gap 1λ1-\lambda, then for any bounded functions fi:x[ai,bi]f_i: x \mapsto [a_i,b_i], the sum of fi(Xi)f_i(X_i) is sub-Gaussian with variance proxy 1+λ1λi(biai)24\frac{1+\lambda}{1-\lambda} \cdot \sum_i \frac{(b_i-a_i)^2}{4}. The counterpart of Hoeffding's inequality immediately follows. Our results assume none of reversibility, countable state space and time-homogeneity of Markov chains. They are optimal in terms of the multiplicative coefficient (1+λ)/(1λ)(1+\lambda)/(1-\lambda), and cover Hoeffding's lemma and inequality for independent random variables as special cases with λ=0\lambda = 0. We illustrate the utility of these results by applying them to six problems in statistics and machine learning. They are linear regression, lasso regression, sparse covariance matrix estimation with Markov-dependent samples; Markov chain Monte Carlo estimation; respondence driven sampling; and multi-armed bandit problems with Markovian rewards.

View on arXiv
Comments on this paper