45
29

Efficiency of adaptive importance sampling

Abstract

The \textit{sampling policy} of stage tt, formally expressed as a probability density function qtq_t, stands for the distribution of the sample (xt,1,,xt,nt)(x_{t,1},\ldots, x_{t,n_t}) generated at tt. From the past samples, some information depending on some \textit{objective} is derived leading eventually to update the sampling policy to qt+1q_{t+1}. This generic approach characterizes \textit{adaptive importance sampling} (AIS) schemes. Each stage tt is formed with two steps : (i) to explore the space with ntn_t points according to qtq_t and (ii) to exploit the current amount of information to update the sampling policy. The very fundamental question raised in the paper concerns the behavior of empirical sums based on AIS. Without making any assumption on the \textit{allocation policy} ntn_t, the theory developed involves no restriction on the split of computational resources between the explore (i) and the exploit (ii) step. It is shown that AIS is efficient : the asymptotic behavior of AIS is the same as some "oracle" strategy that knows the optimal sampling policy from the beginning. From a practical perspective, weighted AIS is introduced, a new method that allows to forget poor samples from early stages.

View on arXiv
Comments on this paper