Safety-Oriented Pruning and Interpretation of Reinforcement Learning
Policies
The European Symposium on Artificial Neural Networks (ESANN), 2024
Helge Spieker
- AAMLOffRL
Main:5 Pages
2 Figures
Bibliography:1 Pages
1 Tables
Appendix:2 Pages
Abstract
Pruning neural networks (NNs) can streamline them but risks removing vital parameters from safe reinforcement learning (RL) policies. We introduce an interpretable RL method called VERINTER, which combines NN pruning with model checking to ensure interpretable RL safety. VERINTER exactly quantifies the effects of pruning and the impact of neural connections on complex safety properties by analyzing changes in safety measurements. This method maintains safety in pruned RL policies and enhances understanding of their safety dynamics, which has proven effective in multiple RL settings.
View on arXivComments on this paper
