42
0

A Large Deviation Approach to Posterior Consistency in Dynamical Systems

Abstract

In this paper, we provide asymptotic results concerning (generalized) Bayesian inference for certain dynamical systems based on a large deviation approach. Given a sequence of observations yy, a class of model processes parameterized by θΘ\theta \in \Theta which can be characterized as a stochastic process XθX^\theta or a measure μθ\mu_\theta, and a loss function LL which measures the error between yy and a realization of XθX^\theta, we specify the generalized posterior distribution πt(θy)\pi_t(\theta \mid y). The goal of this paper is to study the asymptotic behavior of πt(θy)\pi_t(\theta \mid y) as t.t \to \infty. In particular, we state conditions on the model family {μθ}θΘ\{\mu_\theta\}_{\theta \in \Theta} and the loss function LL such that the posterior distribution converges. The two conditions we require are: (1) a conditional large deviation behavior for a single XθX^\theta, and (2) an exponential continuity condition over the model family for the map from the parameter θ\theta to the loss incurred between XθX^\theta and the observation sequence yy. The proposed framework is quite general, we apply it to two very different classes of dynamical systems: continuous time hypermixing processes and Gibbs processes on shifts of finite type. We also show that the generalized posterior distribution concentrates asymptotically on those parameters that minimize the expected loss and a divergence term, hence proving posterior consistency.

View on arXiv
Comments on this paper