ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2011.08434
  4. Cited By
Simple and optimal methods for stochastic variational inequalities, II:
  Markovian noise and policy evaluation in reinforcement learning
v1v2v3v4 (latest)

Simple and optimal methods for stochastic variational inequalities, II: Markovian noise and policy evaluation in reinforcement learning

SIAM Journal on Optimization (SIOPT), 2020
15 November 2020
Georgios Kotsalis
Guanghui Lan
Tianjiao Li
    OffRL
ArXiv (abs)PDFHTMLGithub

Papers citing "Simple and optimal methods for stochastic variational inequalities, II: Markovian noise and policy evaluation in reinforcement learning"

19 / 19 papers shown
Auto-conditioned primal-dual hybrid gradient method and alternating
  direction method of multipliers
Auto-conditioned primal-dual hybrid gradient method and alternating direction method of multipliers
Guanghui Lan
Tianjiao Li
194
4
0
02 Oct 2024
Exploiting Approximate Symmetry for Efficient Multi-Agent Reinforcement
  Learning
Exploiting Approximate Symmetry for Efficient Multi-Agent Reinforcement LearningConference on Learning for Dynamics & Control (L4DC), 2024
Batuhan Yardim
Niao He
AI4CE
298
7
0
27 Aug 2024
Reinforcement Learning for SBM Graphon Games with Re-Sampling
Reinforcement Learning for SBM Graphon Games with Re-Sampling
Peihan Huo
Oscar Peralta
Junyu Guo
Qiaomin Xie
Andreea Minca
189
1
0
25 Oct 2023
A simple uniformly optimal method without line search for convex
  optimization
A simple uniformly optimal method without line search for convex optimization
Tianjiao Li
Guanghui Lan
473
47
0
16 Oct 2023
First-order Policy Optimization for Robust Policy Evaluation
First-order Policy Optimization for Robust Policy Evaluation
Jian Wang
Guanghui Lan
OffRL
207
7
0
29 Jul 2023
Accelerated stochastic approximation with state-dependent noise
Accelerated stochastic approximation with state-dependent noiseMathematical programming (Math. Program.), 2023
Sasila Ilandarideva
A. Juditsky
Guanghui Lan
Tianjiao Li
420
13
0
04 Jul 2023
Last-Iterate Convergence of Adaptive Riemannian Gradient Descent for Equilibrium Computation
Last-Iterate Convergence of Adaptive Riemannian Gradient Descent for Equilibrium Computation
Yong Cai
Michael I. Jordan
Tianyi Lin
Argyris Oikonomou
Emmanouil-Vasileios Vlatakis-Gkaragkounis
371
5
0
29 Jun 2023
Networked Communication for Decentralised Agents in Mean-Field Games
Networked Communication for Decentralised Agents in Mean-Field Games
Patrick Benjamin
Alessandro Abate
FedML
570
2
0
05 Jun 2023
Policy Mirror Descent Inherently Explores Action Space
Policy Mirror Descent Inherently Explores Action SpaceSIAM Journal on Optimization (SIOPT), 2023
Yan Li
Guanghui Lan
OffRL
398
12
0
08 Mar 2023
Policy Mirror Ascent for Efficient and Independent Learning in Mean
  Field Games
Policy Mirror Ascent for Efficient and Independent Learning in Mean Field GamesInternational Conference on Machine Learning (ICML), 2022
Batuhan Yardim
Semih Cayci
Matthieu Geist
Niao He
407
33
0
29 Dec 2022
Smooth Monotone Stochastic Variational Inequalities and Saddle Point
  Problems: A Survey
Smooth Monotone Stochastic Variational Inequalities and Saddle Point Problems: A SurveyEuropean Mathematical Society Magazine (EMS Magazine), 2022
Aleksandr Beznosikov
Boris Polyak
Eduard A. Gorbunov
D. Kovalev
Alexander Gasnikov
408
35
0
29 Aug 2022
Stochastic first-order methods for average-reward Markov decision
  processes
Stochastic first-order methods for average-reward Markov decision processesMathematics of Operations Research (MOR), 2022
Tianjiao Li
Feiyang Wu
Guanghui Lan
589
29
0
11 May 2022
Data Sampling Affects the Complexity of Online SGD over Dependent Data
Data Sampling Affects the Complexity of Online SGD over Dependent DataConference on Uncertainty in Artificial Intelligence (UAI), 2022
Shaocong Ma
Ziyi Chen
Yi Zhou
Kaiyi Ji
Yingbin Liang
339
6
0
31 Mar 2022
Optimal variance-reduced stochastic approximation in Banach spaces
Optimal variance-reduced stochastic approximation in Banach spaces
Wenlong Mou
K. Khamaru
Martin J. Wainwright
Peter L. Bartlett
Sai Li
292
11
0
21 Jan 2022
Block Policy Mirror Descent
Block Policy Mirror Descent
Guanghui Lan
Yan Li
T. Zhao
OffRL
325
11
0
15 Jan 2022
Accelerated and instance-optimal policy evaluation with linear function
  approximation
Accelerated and instance-optimal policy evaluation with linear function approximationSIAM Journal on Mathematics of Data Science (SIMODS), 2021
Tianjiao Li
Guanghui Lan
A. Pananjady
OffRL
259
18
0
24 Dec 2021
PER-ETD: A Polynomially Efficient Emphatic Temporal Difference Learning
  Method
PER-ETD: A Polynomially Efficient Emphatic Temporal Difference Learning Method
Ziwei Guan
Tengyu Xu
Yingbin Liang
250
4
0
13 Oct 2021
Cyclic Coordinate Dual Averaging with Extrapolation
Cyclic Coordinate Dual Averaging with ExtrapolationSIAM Journal on Optimization (SIAM J. Optim.), 2021
Chaobing Song
Jelena Diakonikolas
454
11
0
26 Feb 2021
Policy Mirror Descent for Reinforcement Learning: Linear Convergence,
  New Sampling Complexity, and Generalized Problem Classes
Policy Mirror Descent for Reinforcement Learning: Linear Convergence, New Sampling Complexity, and Generalized Problem ClassesMathematical programming (Math. Program.), 2021
Guanghui Lan
904
176
0
30 Jan 2021
1
Page 1 of 1