18
0

Neuron-level Balance between Stability and Plasticity in Deep Reinforcement Learning

Abstract

In contrast to the human ability to continuously acquire knowledge, agents struggle with the stability-plasticity dilemma in deep reinforcement learning (DRL), which refers to the trade-off between retaining existing skills (stability) and learning new knowledge (plasticity). Current methods focus on balancing these two aspects at the network level, lacking sufficient differentiation and fine-grained control of individual neurons. To overcome this limitation, we propose Neuron-level Balance between Stability and Plasticity (NBSP) method, by taking inspiration from the observation that specific neurons are strongly relevant to task-relevant skills. Specifically, NBSP first (1) defines and identifies RL skill neurons that are crucial for knowledge retention through a goal-oriented method, and then (2) introduces a framework by employing gradient masking and experience replay techniques targeting these neurons to preserve the encoded existing skills while enabling adaptation to new tasks. Numerous experimental results on the Meta-World and Atari benchmarks demonstrate that NBSP significantly outperforms existing approaches in balancing stability and plasticity.

View on arXiv
@article{lan2025_2504.08000,
  title={ Neuron-level Balance between Stability and Plasticity in Deep Reinforcement Learning },
  author={ Jiahua Lan and Sen Zhang and Haixia Pan and Ruijun Liu and Li Shen and Dacheng Tao },
  journal={arXiv preprint arXiv:2504.08000},
  year={ 2025 }
}
Comments on this paper