Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2201.07296
Cited By
Convergence of Policy Gradient for Entropy Regularized MDPs with Neural Network Approximation in the Mean-Field Regime
18 January 2022
B. Kerimkulov
J. Leahy
David Siska
Lukasz Szpruch
Re-assign community
ArXiv
PDF
HTML
Papers citing
"Convergence of Policy Gradient for Entropy Regularized MDPs with Neural Network Approximation in the Mean-Field Regime"
8 / 8 papers shown
Title
Linear convergence of proximal descent schemes on the Wasserstein space
Razvan-Andrei Lascu
Mateusz B. Majka
David Siska
Łukasz Szpruch
72
1
0
22 Nov 2024
A Fisher-Rao gradient flow for entropy-regularised Markov decision processes in Polish spaces
B. Kerimkulov
J. Leahy
David Siska
Lukasz Szpruch
Yufei Zhang
16
7
0
04 Oct 2023
Policy Optimization for Continuous Reinforcement Learning
Hanyang Zhao
Wenpin Tang
D. Yao
OffRL
26
17
0
30 May 2023
Matryoshka Policy Gradient for Entropy-Regularized RL: Convergence and Global Optimality
François Ged
M. H. Veiga
21
0
0
22 Mar 2023
Beyond Exponentially Fast Mixing in Average-Reward Reinforcement Learning via Multi-Level Monte Carlo Actor-Critic
Wesley A. Suttle
Amrit Singh Bedi
Bhrij Patel
Brian M. Sadler
Alec Koppel
Dinesh Manocha
16
13
0
28 Jan 2023
Geometry and convergence of natural policy gradient methods
Johannes Muller
Guido Montúfar
8
9
0
03 Nov 2022
Convergence of policy gradient methods for finite-horizon exploratory linear-quadratic control problems
Michael Giegrich
Christoph Reisinger
Yufei Zhang
16
11
0
01 Nov 2022
Linear convergence of a policy gradient method for some finite horizon continuous time control problems
C. Reisinger
Wolfgang Stockinger
Yufei Zhang
16
5
0
22 Mar 2022
1