Optimal Uniform OPE and Model-based Offline Reinforcement Learning in
Time-Homogeneous, Reward-Free and Task-Agnostic Settings
- OffRL
This work studies the statistical limits of uniform convergence for offline policy evaluation (OPE) problems with model-based methods (for finite horizon MDP) and provides a unified view towards optimal learning for several well-motivated offline tasks. Uniform OPE (initiated by \citet{yin2021near}) is a stronger measure than the point-wise (fixed policy) OPE and ensures offline policy learning when contains all policies (global policy class). In this paper, we establish an lower bound (over model-based family) for the global uniform OPE, where is the minimal state-action probability induced by the behavior policy. Next, our main result establishes an episode complexity of for \emph{local} uniform convergence that applies to all \emph{near-empirically optimal} policies for the MDPs with \emph{stationary} transition. This result implies the optimal sample complexity for offline learning and separates the local uniform OPE from the global case due to the extra factor. Paramountly, the model-based method combining with our new analysis technique (singleton absorbing MDP) can be adapted to the new settings: offline task-agnostic and the offline reward-free with optimal complexity ( is the number of tasks) and respectively, which provides a unified framework for simultaneously solving different offline RL problems.
View on arXiv