0
0

A Robust Model-Based Approach for Continuous-Time Policy Evaluation with Unknown Lévy Process Dynamics

Qihao Ye
Xiaochuan Tian
Yuhua Zhu
Abstract

This paper develops a model-based framework for continuous-time policy evaluation (CTPE) in reinforcement learning, incorporating both Brownian and Lévy noise to model stochastic dynamics influenced by rare and extreme events. Our approach formulates the policy evaluation problem as solving a partial integro-differential equation (PIDE) for the value function with unknown coefficients. A key challenge in this setting is accurately recovering the unknown coefficients in the stochastic dynamics, particularly when driven by Lévy processes with heavy tail effects. To address this, we propose a robust numerical approach that effectively handles both unbiased and censored trajectory datasets. This method combines maximum likelihood estimation with an iterative tail correction mechanism, improving the stability and accuracy of coefficient recovery. Additionally, we establish a theoretical bound for the policy evaluation error based on coefficient recovery error. Through numerical experiments, we demonstrate the effectiveness and robustness of our method in recovering heavy-tailed Lévy dynamics and verify the theoretical error analysis in policy evaluation.

View on arXiv
@article{ye2025_2504.01482,
  title={ A Robust Model-Based Approach for Continuous-Time Policy Evaluation with Unknown Lévy Process Dynamics },
  author={ Qihao Ye and Xiaochuan Tian and Yuhua Zhu },
  journal={arXiv preprint arXiv:2504.01482},
  year={ 2025 }
}
Comments on this paper