ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2203.13935
  4. Cited By
Offline Reinforcement Learning Under Value and Density-Ratio
  Realizability: The Power of Gaps

Offline Reinforcement Learning Under Value and Density-Ratio Realizability: The Power of Gaps

25 March 2022
Jinglin Chen
Nan Jiang
    OffRL
ArXivPDFHTML

Papers citing "Offline Reinforcement Learning Under Value and Density-Ratio Realizability: The Power of Gaps"

24 / 24 papers shown
Title
Online Reinforcement Learning with Passive Memory
Online Reinforcement Learning with Passive Memory
Anay Pattanaik
Lav R. Varshney
CLL
OffRL
15
0
0
18 Oct 2024
The Role of Inherent Bellman Error in Offline Reinforcement Learning
  with Linear Function Approximation
The Role of Inherent Bellman Error in Offline Reinforcement Learning with Linear Function Approximation
Noah Golowich
Ankur Moitra
OffRL
21
2
0
17 Jun 2024
A Unified Linear Programming Framework for Offline Reward Learning from
  Human Demonstrations and Feedback
A Unified Linear Programming Framework for Offline Reward Learning from Human Demonstrations and Feedback
Kihyun Kim
Jiawei Zhang
Asuman Ozdaglar
P. Parrilo
OffRL
33
1
0
20 May 2024
On Sample-Efficient Offline Reinforcement Learning: Data Diversity,
  Posterior Sampling, and Beyond
On Sample-Efficient Offline Reinforcement Learning: Data Diversity, Posterior Sampling, and Beyond
Thanh Nguyen-Tang
Raman Arora
OffRL
15
3
0
06 Jan 2024
Bi-Level Offline Policy Optimization with Limited Exploration
Bi-Level Offline Policy Optimization with Limited Exploration
Wenzhuo Zhou
OffRL
34
4
0
10 Oct 2023
Stackelberg Batch Policy Learning
Stackelberg Batch Policy Learning
Wenzhuo Zhou
Annie Qu
OffRL
14
0
0
28 Sep 2023
Distributional Shift-Aware Off-Policy Interval Estimation: A Unified
  Error Quantification Framework
Distributional Shift-Aware Off-Policy Interval Estimation: A Unified Error Quantification Framework
Wenzhuo Zhou
Yuhan Li
Ruoqing Zhu
Annie Qu
OffRL
8
4
0
23 Sep 2023
Bayesian Safe Policy Learning with Chance Constrained Optimization:
  Application to Military Security Assessment during the Vietnam War
Bayesian Safe Policy Learning with Chance Constrained Optimization: Application to Military Security Assessment during the Vietnam War
Zeyang Jia
Eli Ben-Michael
Kosuke Imai
11
4
0
17 Jul 2023
A Primal-Dual-Critic Algorithm for Offline Constrained Reinforcement
  Learning
A Primal-Dual-Critic Algorithm for Offline Constrained Reinforcement Learning
Kihyuk Hong
Yuhang Li
Ambuj Tewari
OffRL
13
7
0
13 Jun 2023
Offline Reinforcement Learning with Additional Covering Distributions
Offline Reinforcement Learning with Additional Covering Distributions
Chenjie Mao
OffRL
10
0
0
22 May 2023
VIPeR: Provably Efficient Algorithm for Offline RL with Neural Function
  Approximation
VIPeR: Provably Efficient Algorithm for Offline RL with Neural Function Approximation
Thanh Nguyen-Tang
R. Arora
OffRL
22
5
0
24 Feb 2023
Robust Knowledge Transfer in Tiered Reinforcement Learning
Robust Knowledge Transfer in Tiered Reinforcement Learning
Jiawei Huang
Niao He
OffRL
8
1
0
10 Feb 2023
Offline Minimax Soft-Q-learning Under Realizability and Partial Coverage
Offline Minimax Soft-Q-learning Under Realizability and Partial Coverage
Masatoshi Uehara
Nathan Kallus
Jason D. Lee
Wen Sun
OffRL
13
5
0
05 Feb 2023
Reinforcement Learning in Low-Rank MDPs with Density Features
Reinforcement Learning in Low-Rank MDPs with Density Features
Audrey Huang
Jinglin Chen
Nan Jiang
OffRL
4
14
0
04 Feb 2023
Importance Weighted Actor-Critic for Optimal Conservative Offline
  Reinforcement Learning
Importance Weighted Actor-Critic for Optimal Conservative Offline Reinforcement Learning
Hanlin Zhu
Paria Rashidinejad
Jiantao Jiao
OffRL
12
15
0
30 Jan 2023
Policy learning "without'' overlap: Pessimism and generalized empirical
  Bernstein's inequality
Policy learning "without'' overlap: Pessimism and generalized empirical Bernstein's inequality
Ying Jin
Zhimei Ren
Zhuoran Yang
Zhaoran Wang
OffRL
14
25
0
19 Dec 2022
Scaling Marginalized Importance Sampling to High-Dimensional
  State-Spaces via State Abstraction
Scaling Marginalized Importance Sampling to High-Dimensional State-Spaces via State Abstraction
Brahma S. Pavse
Josiah P. Hanna
OffRL
19
7
0
14 Dec 2022
Leveraging Offline Data in Online Reinforcement Learning
Leveraging Offline Data in Online Reinforcement Learning
Andrew Wagenmaker
Aldo Pacchiano
OffRL
OnRL
11
36
0
09 Nov 2022
Optimal Conservative Offline RL with General Function Approximation via
  Augmented Lagrangian
Optimal Conservative Offline RL with General Function Approximation via Augmented Lagrangian
Paria Rashidinejad
Hanlin Zhu
Kunhe Yang
Stuart J. Russell
Jiantao Jiao
OffRL
33
26
0
01 Nov 2022
Hybrid RL: Using Both Offline and Online Data Can Make RL Efficient
Hybrid RL: Using Both Offline and Online Data Can Make RL Efficient
Yuda Song
Yi Zhou
Ayush Sekhari
J. Andrew Bagnell
A. Krishnamurthy
Wen Sun
OffRL
OnRL
25
89
0
13 Oct 2022
On the Statistical Efficiency of Reward-Free Exploration in Non-Linear
  RL
On the Statistical Efficiency of Reward-Free Exploration in Non-Linear RL
Jinglin Chen
Aditya Modi
A. Krishnamurthy
Nan Jiang
Alekh Agarwal
19
25
0
21 Jun 2022
Pessimism for Offline Linear Contextual Bandits using $\ell_p$
  Confidence Sets
Pessimism for Offline Linear Contextual Bandits using ℓp\ell_pℓp​ Confidence Sets
Gen Li
Cong Ma
Nathan Srebro
OffRL
10
11
0
21 May 2022
Sample Complexity of Offline Reinforcement Learning with Deep ReLU
  Networks
Sample Complexity of Offline Reinforcement Learning with Deep ReLU Networks
Thanh Nguyen-Tang
Sunil R. Gupta
Hung The Tran
Svetha Venkatesh
OffRL
44
7
0
11 Mar 2021
Model-free Representation Learning and Exploration in Low-rank MDPs
Model-free Representation Learning and Exploration in Low-rank MDPs
Aditya Modi
Jinglin Chen
A. Krishnamurthy
Nan Jiang
Alekh Agarwal
OffRL
98
78
0
14 Feb 2021
1