Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2301.13139
Cited By
A Novel Framework for Policy Mirror Descent with General Parameterization and Linear Convergence
30 January 2023
Carlo Alfano
Rui Yuan
Patrick Rebeschini
Re-assign community
ArXiv
PDF
HTML
Papers citing
"A Novel Framework for Policy Mirror Descent with General Parameterization and Linear Convergence"
10 / 10 papers shown
Title
Policy Mirror Descent Inherently Explores Action Space
Yan Li
Guanghui Lan
17
6
0
08 Mar 2023
Policy Mirror Ascent for Efficient and Independent Learning in Mean Field Games
Batuhan Yardim
Semih Cayci
M. Geist
Niao He
20
19
0
29 Dec 2022
Linear Convergence for Natural Policy Gradient with Log-linear Policy Parametrization
Carlo Alfano
Patrick Rebeschini
30
13
0
30 Sep 2022
Stochastic Second-Order Methods Improve Best-Known Sample Complexity of SGD for Gradient-Dominated Function
Saeed Masiha
Saber Salehkaleybar
Niao He
Negar Kiyavash
Patrick Thiran
42
17
0
25 May 2022
Actor-critic is implicitly biased towards high entropy optimal policies
Yuzheng Hu
Ziwei Ji
Matus Telgarsky
15
11
0
21 Oct 2021
Finite-Sample Analysis of Off-Policy Natural Actor-Critic Algorithm
S. Khodadadian
Zaiwei Chen
S. T. Maguluri
CML
18
25
0
18 Feb 2021
On the Convergence and Sample Efficiency of Variance-Reduced Policy Gradient Method
Junyu Zhang
Chengzhuo Ni
Zheng Yu
Csaba Szepesvári
Mengdi Wang
22
57
0
17 Feb 2021
Policy Mirror Descent for Reinforcement Learning: Linear Convergence, New Sampling Complexity, and Generalized Problem Classes
Guanghui Lan
28
120
0
30 Jan 2021
On Linear Convergence of Policy Gradient Methods for Finite MDPs
Jalaj Bhandari
Daniel Russo
15
55
0
21 Jul 2020
Offline Reinforcement Learning: Tutorial, Review, and Perspectives on Open Problems
Sergey Levine
Aviral Kumar
George Tucker
Justin Fu
GP
240
1,662
0
04 May 2020
1