ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2007.04803
49
4
v1v2v3v4v5v6v7v8v9 (latest)

A Global Stochastic Optimization Particle Filter Algorithm

9 July 2020
Mathieu Gerber
Randal Douc
ArXiv (abs)PDFHTML
Abstract

We introduce a new algorithm to learn on the fly the parameter value θ⋆:=argmaxθ∈ΘE[log⁡fθ(Y0)]\theta_\star:=\mathrm{argmax}_{\theta\in\Theta}\mathbb{E}[\log f_\theta(Y_0)]θ⋆​:=argmaxθ∈Θ​E[logfθ​(Y0​)] from a sequence (Yt)t≥1(Y_t)_{t\geq 1}(Yt​)t≥1​ of independent copies of Y0Y_0Y0​, with {fθ, θ∈Θ⊆Rd}\{f_\theta,\,\theta\in\Theta\subseteq\mathbb{R}^d\}{fθ​,θ∈Θ⊆Rd} a parametric model. The main idea of the proposed approach is to define a sequence (π~t)t≥1(\tilde{\pi}_t)_{t\geq 1}(π~t​)t≥1​ of probability distributions on Θ\ThetaΘ which (i) is shown to concentrate on θ⋆\theta_\starθ⋆​ as t→∞t\rightarrow\inftyt→∞ and (ii) can be estimated in an online fashion by means of a standard particle filter (PF) algorithm. The sequence (π~t)t≥1(\tilde{\pi}_t)_{t\geq 1}(π~t​)t≥1​ depends on a learning rate ht→0h_t\rightarrow 0ht​→0, with the slower hth_tht​ converges to zero the greater is the ability of the PF approximation π~tN\tilde{\pi}_t^Nπ~tN​ of π~t\tilde{\pi}_tπ~t​ to escape from a local optimum of the objective function, but the slower is the rate at which π~t\tilde{\pi}_tπ~t​ concentrates on θ⋆\theta_\starθ⋆​. To conciliate ability to escape from a local optimum and fast convergence towards θ⋆\theta_\starθ⋆​ we exploit the acceleration property of averaging, well-known in the stochastic gradient descent literature, by letting θˉtN:=t−1∑s=1t∫Θθ π~sN(dθ)\bar{\theta}_t^N:=t^{-1}\sum_{s=1}^t \int_{\Theta}\theta\ \tilde{\pi}_s^N(\mathrm{d} \theta)θˉtN​:=t−1∑s=1t​∫Θ​θ π~sN​(dθ) be the proposed estimator of θ⋆\theta_\starθ⋆​. Our numerical experiments suggest that θˉtN\bar{\theta}_t^NθˉtN​ converges to θ⋆\theta_\starθ⋆​ at the optimal t−1/2t^{-1/2}t−1/2 rate in challenging models and in situations where π~tN\tilde{\pi}_t^Nπ~tN​ concentrates on this parameter value at a slower rate. We illustrate the practical usefulness of the proposed optimization algorithm for online parameter learning and for computing the maximum likelihood estimator.

View on arXiv
Comments on this paper