ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2012.08005
  4. Cited By
Exponential Lower Bounds for Batch Reinforcement Learning: Batch RL can
  be Exponentially Harder than Online RL
v1v2v3v4 (latest)

Exponential Lower Bounds for Batch Reinforcement Learning: Batch RL can be Exponentially Harder than Online RL

International Conference on Machine Learning (ICML), 2020
14 December 2020
Andrea Zanette
    OffRL
ArXiv (abs)PDFHTML

Papers citing "Exponential Lower Bounds for Batch Reinforcement Learning: Batch RL can be Exponentially Harder than Online RL"

50 / 53 papers shown
Offline Reinforcement Learning in Large State Spaces: Algorithms and Guarantees
Offline Reinforcement Learning in Large State Spaces: Algorithms and Guarantees
Nan Jiang
Tengyang Xie
OffRL
221
14
0
05 Oct 2025
A Tutorial: An Intuitive Explanation of Offline Reinforcement Learning Theory
A Tutorial: An Intuitive Explanation of Offline Reinforcement Learning Theory
Fengdi Che
OffRL
172
0
0
11 Aug 2025
Central Limit Theorems for Transition Probabilities of Controlled Markov Chains
Central Limit Theorems for Transition Probabilities of Controlled Markov Chains
Ziwei Su
Imon Banerjee
Diego Klabjan
OffRL
218
0
0
02 Aug 2025
The Role of Inherent Bellman Error in Offline Reinforcement Learning
  with Linear Function Approximation
The Role of Inherent Bellman Error in Offline Reinforcement Learning with Linear Function Approximation
Noah Golowich
Ankur Moitra
OffRL
342
3
0
17 Jun 2024
Trajectory Data Suffices for Statistically Efficient Learning in Offline
  RL with Linear $q^π$-Realizability and Concentrability
Trajectory Data Suffices for Statistically Efficient Learning in Offline RL with Linear qπq^πqπ-Realizability and Concentrability
Volodymyr Tkachuk
Gellert Weisz
Csaba Szepesvári
OffRL
221
3
0
27 May 2024
Experiment Planning with Function Approximation
Experiment Planning with Function ApproximationNeural Information Processing Systems (NeurIPS), 2024
Aldo Pacchiano
Jonathan Lee
Emma Brunskill
OffRL
219
6
0
10 Jan 2024
Butterfly Effects of SGD Noise: Error Amplification in Behavior Cloning
  and Autoregression
Butterfly Effects of SGD Noise: Error Amplification in Behavior Cloning and AutoregressionInternational Conference on Learning Representations (ICLR), 2023
Adam Block
Dylan J. Foster
Akshay Krishnamurthy
Max Simchowitz
Cyril Zhang
302
11
0
17 Oct 2023
When is Agnostic Reinforcement Learning Statistically Tractable?
When is Agnostic Reinforcement Learning Statistically Tractable?Neural Information Processing Systems (NeurIPS), 2023
Zeyu Jia
Gene Li
Alexander Rakhlin
Ayush Sekhari
Nathan Srebro
OffRL
345
7
0
09 Oct 2023
Sample-Efficiency in Multi-Batch Reinforcement Learning: The Need for
  Dimension-Dependent Adaptivity
Sample-Efficiency in Multi-Batch Reinforcement Learning: The Need for Dimension-Dependent AdaptivityInternational Conference on Learning Representations (ICLR), 2023
Emmeran Johnson
Ciara Pike-Burke
Patrick Rebeschini
OffRL
343
2
0
02 Oct 2023
What can online reinforcement learning with function approximation
  benefit from general coverage conditions?
What can online reinforcement learning with function approximation benefit from general coverage conditions?International Conference on Machine Learning (ICML), 2023
Fanghui Liu
Luca Viano
Volkan Cevher
OffRL
282
6
0
25 Apr 2023
A Unified Framework of Policy Learning for Contextual Bandit with
  Confounding Bias and Missing Observations
A Unified Framework of Policy Learning for Contextual Bandit with Confounding Bias and Missing Observations
Siyu Chen
Yitan Wang
Zhaoran Wang
Zhuoran Yang
OffRL
234
3
0
20 Mar 2023
VIPeR: Provably Efficient Algorithm for Offline RL with Neural Function
  Approximation
VIPeR: Provably Efficient Algorithm for Offline RL with Neural Function ApproximationInternational Conference on Learning Representations (ICLR), 2023
Thanh Nguyen-Tang
R. Arora
OffRL
304
6
0
24 Feb 2023
Model-based Offline Reinforcement Learning with Local Misspecification
Model-based Offline Reinforcement Learning with Local MisspecificationAAAI Conference on Artificial Intelligence (AAAI), 2023
Kefan Dong
Yannis Flet-Berliac
Allen Nie
Emma Brunskill
OffRL
240
6
0
26 Jan 2023
Data-Driven Offline Decision-Making via Invariant Representation
  Learning
Data-Driven Offline Decision-Making via Invariant Representation LearningNeural Information Processing Systems (NeurIPS), 2022
Qi
Yi-Hsun Su
Aviral Kumar
Sergey Levine
OffRL
322
35
0
21 Nov 2022
When is Realizability Sufficient for Off-Policy Reinforcement Learning?
When is Realizability Sufficient for Off-Policy Reinforcement Learning?International Conference on Machine Learning (ICML), 2022
Andrea Zanette
OffRL
347
16
0
10 Nov 2022
Oracle Inequalities for Model Selection in Offline Reinforcement
  Learning
Oracle Inequalities for Model Selection in Offline Reinforcement LearningNeural Information Processing Systems (NeurIPS), 2022
Jonathan Lee
George Tucker
Ofir Nachum
Bo Dai
Emma Brunskill
OffRL
364
14
0
03 Nov 2022
Optimal Conservative Offline RL with General Function Approximation via
  Augmented Lagrangian
Optimal Conservative Offline RL with General Function Approximation via Augmented LagrangianInternational Conference on Learning Representations (ICLR), 2022
Paria Rashidinejad
Hanlin Zhu
Kunhe Yang
Stuart J. Russell
Jiantao Jiao
OffRL
455
34
0
01 Nov 2022
Reliable Conditioning of Behavioral Cloning for Offline Reinforcement
  Learning
Reliable Conditioning of Behavioral Cloning for Offline Reinforcement Learning
Tung Nguyen
Qinqing Zheng
Aditya Grover
OffRL
340
7
0
11 Oct 2022
Distributionally Robust Offline Reinforcement Learning with Linear
  Function Approximation
Distributionally Robust Offline Reinforcement Learning with Linear Function Approximation
Xiaoteng Ma
Zhipeng Liang
Jose H. Blanchet
MingWen Liu
Li Xia
Jiheng Zhang
Qianchuan Zhao
Zhengyuan Zhou
OODOffRL
355
33
0
14 Sep 2022
Provably Efficient Offline Reinforcement Learning with Trajectory-Wise
  Reward
Provably Efficient Offline Reinforcement Learning with Trajectory-Wise RewardIEEE Transactions on Information Theory (IEEE Trans. Inf. Theory), 2022
Tengyu Xu
Yue Wang
Shaofeng Zou
Yingbin Liang
OffRL
262
16
0
13 Jun 2022
Offline Reinforcement Learning with Differential Privacy
Offline Reinforcement Learning with Differential PrivacyNeural Information Processing Systems (NeurIPS), 2022
Dan Qiao
Yu Wang
OffRL
417
29
0
02 Jun 2022
Stabilizing Q-learning with Linear Architectures for Provably Efficient
  Learning
Stabilizing Q-learning with Linear Architectures for Provably Efficient LearningInternational Conference on Machine Learning (ICML), 2022
Andrea Zanette
Martin J. Wainwright
OOD
300
5
0
01 Jun 2022
Pessimism in the Face of Confounders: Provably Efficient Offline
  Reinforcement Learning in Partially Observable Markov Decision Processes
Pessimism in the Face of Confounders: Provably Efficient Offline Reinforcement Learning in Partially Observable Markov Decision ProcessesInternational Conference on Learning Representations (ICLR), 2022
Miao Lu
Yifei Min
Zhaoran Wang
Zhuoran Yang
OffRL
414
26
0
26 May 2022
Pessimism for Offline Linear Contextual Bandits using $\ell_p$
  Confidence Sets
Pessimism for Offline Linear Contextual Bandits using ℓp\ell_pℓp​ Confidence SetsNeural Information Processing Systems (NeurIPS), 2022
Gen Li
Cong Ma
Nathan Srebro
OffRL
323
19
0
21 May 2022
When Should We Prefer Offline Reinforcement Learning Over Behavioral
  Cloning?
When Should We Prefer Offline Reinforcement Learning Over Behavioral Cloning?
Aviral Kumar
Joey Hong
Anika Singh
Sergey Levine
OffRL
317
100
0
12 Apr 2022
Bellman Residual Orthogonalization for Offline Reinforcement Learning
Bellman Residual Orthogonalization for Offline Reinforcement LearningNeural Information Processing Systems (NeurIPS), 2022
Andrea Zanette
Martin J. Wainwright
OffRL
502
12
0
24 Mar 2022
Near-optimal Offline Reinforcement Learning with Linear Representation:
  Leveraging Variance Information with Pessimism
Near-optimal Offline Reinforcement Learning with Linear Representation: Leveraging Variance Information with PessimismInternational Conference on Learning Representations (ICLR), 2022
Ming Yin
Yaqi Duan
Mengdi Wang
Yu Wang
OffRL
285
71
0
11 Mar 2022
A Complete Characterization of Linear Estimators for Offline Policy
  Evaluation
A Complete Characterization of Linear Estimators for Offline Policy EvaluationJournal of machine learning research (JMLR), 2022
Juan C. Perdomo
A. Krishnamurthy
Peter L. Bartlett
Sham Kakade
OffRL
290
5
0
08 Mar 2022
Offline Reinforcement Learning with Realizability and Single-policy
  Concentrability
Offline Reinforcement Learning with Realizability and Single-policy ConcentrabilityAnnual Conference Computational Learning Theory (COLT), 2022
Wenhao Zhan
Baihe Huang
Audrey Huang
Nan Jiang
Jason D. Lee
OffRL
654
122
0
09 Feb 2022
Why Should I Trust You, Bellman? The Bellman Error is a Poor Replacement
  for Value Error
Why Should I Trust You, Bellman? The Bellman Error is a Poor Replacement for Value ErrorInternational Conference on Machine Learning (ICML), 2022
Scott Fujimoto
David Meger
Doina Precup
Ofir Nachum
S. Gu
381
41
0
28 Jan 2022
Accelerated and instance-optimal policy evaluation with linear function
  approximation
Accelerated and instance-optimal policy evaluation with linear function approximationSIAM Journal on Mathematics of Data Science (SIMODS), 2021
Tianjiao Li
Guanghui Lan
A. Pananjady
OffRL
236
16
0
24 Dec 2021
DR3: Value-Based Deep Reinforcement Learning Requires Explicit
  Regularization
DR3: Value-Based Deep Reinforcement Learning Requires Explicit Regularization
Aviral Kumar
Rishabh Agarwal
Tengyu Ma
Aaron Courville
George Tucker
Sergey Levine
OffRL
231
83
0
09 Dec 2021
The Impact of Data Distribution on Q-learning with Function
  Approximation
The Impact of Data Distribution on Q-learning with Function Approximation
Pedro P. Santos
Diogo S. Carvalho
Alberto Sardinha
Francisco S. Melo
OffRL
265
3
0
23 Nov 2021
Offline Reinforcement Learning: Fundamental Barriers for Value Function
  Approximation
Offline Reinforcement Learning: Fundamental Barriers for Value Function ApproximationAnnual Conference Computational Learning Theory (COLT), 2021
Dylan J. Foster
A. Krishnamurthy
D. Simchi-Levi
Yunzong Xu
OffRL
334
74
0
21 Nov 2021
Exploiting Action Impact Regularity and Exogenous State Variables for
  Offline Reinforcement Learning
Exploiting Action Impact Regularity and Exogenous State Variables for Offline Reinforcement LearningJournal of Artificial Intelligence Research (JAIR), 2021
Vincent Liu
James Wright
Martha White
OffRL
342
2
0
15 Nov 2021
Towards Instance-Optimal Offline Reinforcement Learning with Pessimism
Towards Instance-Optimal Offline Reinforcement Learning with Pessimism
Ming Yin
Yu Wang
OffRL
318
88
0
17 Oct 2021
Provable Benefits of Actor-Critic Methods for Offline Reinforcement
  Learning
Provable Benefits of Actor-Critic Methods for Offline Reinforcement Learning
Andrea Zanette
Martin J. Wainwright
Emma Brunskill
OffRL
340
132
0
19 Aug 2021
Provably Efficient Generative Adversarial Imitation Learning for Online
  and Offline Setting with Linear Function Approximation
Provably Efficient Generative Adversarial Imitation Learning for Online and Offline Setting with Linear Function Approximation
Zhihan Liu
Yufeng Zhang
Zuyue Fu
Zhuoran Yang
Zhaoran Wang
OffRL
150
8
0
19 Aug 2021
Agnostic Reinforcement Learning with Low-Rank MDPs and Rich Observations
Agnostic Reinforcement Learning with Low-Rank MDPs and Rich ObservationsNeural Information Processing Systems (NeurIPS), 2021
Christoph Dann
Yishay Mansour
M. Mohri
Ayush Sekhari
Karthik Sridharan
OffRL
138
13
0
22 Jun 2021
The Curse of Passive Data Collection in Batch Reinforcement Learning
The Curse of Passive Data Collection in Batch Reinforcement Learning
Chenjun Xiao
Ilbin Lee
Bo Dai
Dale Schuurmans
Csaba Szepesvári
OffRL
255
1
0
18 Jun 2021
Offline RL Without Off-Policy Evaluation
Offline RL Without Off-Policy Evaluation
David Brandfonbrener
William F. Whitney
Rajesh Ranganath
Joan Bruna
OffRL
350
191
0
16 Jun 2021
Bellman-consistent Pessimism for Offline Reinforcement Learning
Bellman-consistent Pessimism for Offline Reinforcement LearningNeural Information Processing Systems (NeurIPS), 2021
Tengyang Xie
Ching-An Cheng
Nan Jiang
Paul Mineiro
Alekh Agarwal
OffRLLRM
755
313
0
13 Jun 2021
Mitigating Covariate Shift in Imitation Learning via Offline Data
  Without Great Coverage
Mitigating Covariate Shift in Imitation Learning via Offline Data Without Great Coverage
Jonathan D. Chang
Masatoshi Uehara
Dhruv Sreenivas
Rahul Kidambi
Wen Sun
OffRL
357
37
0
06 Jun 2021
Optimal Uniform OPE and Model-based Offline Reinforcement Learning in
  Time-Homogeneous, Reward-Free and Task-Agnostic Settings
Optimal Uniform OPE and Model-based Offline Reinforcement Learning in Time-Homogeneous, Reward-Free and Task-Agnostic SettingsNeural Information Processing Systems (NeurIPS), 2021
Ming Yin
Yu Wang
OffRL
300
19
0
13 May 2021
Cautiously Optimistic Policy Optimization and Exploration with Linear
  Function Approximation
Cautiously Optimistic Policy Optimization and Exploration with Linear Function ApproximationAnnual Conference Computational Learning Theory (COLT), 2021
Andrea Zanette
Ching-An Cheng
Alekh Agarwal
343
58
0
24 Mar 2021
An Exponential Lower Bound for Linearly-Realizable MDPs with Constant
  Suboptimality Gap
An Exponential Lower Bound for Linearly-Realizable MDPs with Constant Suboptimality GapNeural Information Processing Systems (NeurIPS), 2021
Yuanhao Wang
Ruosong Wang
Sham Kakade
OffRL
418
48
0
23 Mar 2021
Bridging Offline Reinforcement Learning and Imitation Learning: A Tale
  of Pessimism
Bridging Offline Reinforcement Learning and Imitation Learning: A Tale of PessimismIEEE Transactions on Information Theory (IEEE Trans. Inf. Theory), 2021
Paria Rashidinejad
Banghua Zhu
Cong Ma
Jiantao Jiao
Stuart J. Russell
OffRL
866
324
0
22 Mar 2021
Infinite-Horizon Offline Reinforcement Learning with Linear Function
  Approximation: Curse of Dimensionality and Algorithm
Infinite-Horizon Offline Reinforcement Learning with Linear Function Approximation: Curse of Dimensionality and Algorithm
Lin Chen
B. Scherrer
Peter L. Bartlett
OffRL
353
16
0
17 Mar 2021
Instabilities of Offline RL with Pre-Trained Neural Representation
Instabilities of Offline RL with Pre-Trained Neural RepresentationInternational Conference on Machine Learning (ICML), 2021
Ruosong Wang
Yifan Wu
Ruslan Salakhutdinov
Sham Kakade
OffRL
369
45
0
08 Mar 2021
Uncertainty Estimation Using Riemannian Model Dynamics for Offline
  Reinforcement Learning
Uncertainty Estimation Using Riemannian Model Dynamics for Offline Reinforcement LearningNeural Information Processing Systems (NeurIPS), 2021
Guy Tennenholtz
Shie Mannor
OffRL
245
15
0
22 Feb 2021
12
Next
Page 1 of 2