27
4

Integrating Reinforcement Learning and Model Predictive Control with Applications to Microgrids

Abstract

This work proposes an approach that integrates reinforcement learning and model predictive control (MPC) to solve finite-horizon optimal control problems in mixed-logical dynamical systems efficiently. Optimization-based control of such systems with discrete and continuous decision variables entails the online solution of mixed-integer linear programs, which suffer from the curse of dimensionality. Our approach aims to mitigate this issue by decoupling the decision on the discrete variables from the decision on the continuous variables. In the proposed approach, reinforcement learning determines the discrete decision variables and simplifies the online optimization problem of the MPC controller from a mixed-integer linear program to a linear program, significantly reducing the computational time. A fundamental contribution of this work is the definition of the decoupled Q-function, which plays a crucial role in making the learning problem tractable in a combinatorial action space. We motivate the use of recurrent neural networks to approximate the decoupled Q-function and show how they can be employed in a reinforcement learning setting. Simulation experiments on a microgrid system using real-world data demonstrate that the proposed method substantially reduces the online computation time of MPC while maintaining high feasibility and low suboptimality.

View on arXiv
@article{silva2025_2409.11267,
  title={ Integrating Reinforcement Learning and Model Predictive Control with Applications to Microgrids },
  author={ Caio Fabio Oliveira da Silva and Azita Dabiri and Bart De Schutter },
  journal={arXiv preprint arXiv:2409.11267},
  year={ 2025 }
}
Comments on this paper