Tail Distribution of Regret in Optimistic Reinforcement Learning
We derive instance-dependent tail bounds for the regret of optimism-based reinforcement learning in finite-horizon tabular Markov decision processes with unknown transition dynamics. We first study a UCBVI-type (model-based) algorithm and characterize the tail distribution of the cumulative regret over episodes via explicit bounds on , going beyond analyses limited to or a single high-probability quantile. We analyze two natural exploration-bonus schedules for UCBVI: (i) a -dependent scheme that explicitly incorporates the total number of episodes , and (ii) a -independent (anytime) scheme that depends only on the current episode index. We then complement the model-based results with an analysis of optimistic Q-learning (model-free) under a -dependent bonus schedule.Across both the model-based and model-free settings, we obtain upper bounds on with a distinctive two-regime structure: a sub-Gaussian tail starting from an instance-dependent scale up to a transition threshold, followed by a sub-Weibull tail beyond that point. We further derive corresponding instance-dependent bounds on the expected regret . The proposed algorithms depend on a tuning parameter , which balances the expected regret and the range over which the regret exhibits sub-Gaussian decay.
View on arXiv