ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1811.06225
  4. Cited By
Reward-estimation variance elimination in sequential decision processes

Reward-estimation variance elimination in sequential decision processes

15 November 2018
S. Pankov
ArXivPDFHTML

Papers citing "Reward-estimation variance elimination in sequential decision processes"

5 / 5 papers shown
Title
Multi-Fidelity Policy Gradient Algorithms
Multi-Fidelity Policy Gradient Algorithms
Xinjie Liu
Cyrus Neary
Kushagra Gupta
Christian Ellis
Ufuk Topcu
David Fridovich-Keil
OffRL
256
0
0
07 Mar 2025
Dealing with the Unknown: Pessimistic Offline Reinforcement Learning
Dealing with the Unknown: Pessimistic Offline Reinforcement Learning
Jinning Li
Chen Tang
Masayoshi Tomizuka
Wei Zhan
OffRL
21
21
0
09 Nov 2021
Coordinate-wise Control Variates for Deep Policy Gradients
Coordinate-wise Control Variates for Deep Policy Gradients
Yuanyi Zhong
Yuanshuo Zhou
Jian-wei Peng
BDL
32
1
0
11 Jul 2021
Offline Reinforcement Learning: Tutorial, Review, and Perspectives on
  Open Problems
Offline Reinforcement Learning: Tutorial, Review, and Perspectives on Open Problems
Sergey Levine
Aviral Kumar
George Tucker
Justin Fu
OffRL
GP
343
1,968
0
04 May 2020
Trajectory-wise Control Variates for Variance Reduction in Policy
  Gradient Methods
Trajectory-wise Control Variates for Variance Reduction in Policy Gradient Methods
Ching-An Cheng
Xinyan Yan
Byron Boots
30
22
0
08 Aug 2019
1