ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2505.24061
15
0

Measure gradients, not activations! Enhancing neuronal activity in deep reinforcement learning

29 May 2025
Jiashun Liu
Zihao Wu
J. Obando-Ceron
Pablo Samuel Castro
Aaron Courville
L. Pan
ArXiv (abs)PDFHTML
Main:11 Pages
15 Figures
Bibliography:5 Pages
8 Tables
Appendix:4 Pages
Abstract

Deep reinforcement learning (RL) agents frequently suffer from neuronal activity loss, which impairs their ability to adapt to new data and learn continually. A common method to quantify and address this issue is the tau-dormant neuron ratio, which uses activation statistics to measure the expressive ability of neurons. While effective for simple MLP-based agents, this approach loses statistical power in more complex architectures. To address this, we argue that in advanced RL agents, maintaining a neuron's learning capacity, its ability to adapt via gradient updates, is more critical than preserving its expressive ability. Based on this insight, we shift the statistical objective from activations to gradients, and introduce GraMa (Gradient Magnitude Neural Activity Metric), a lightweight, architecture-agnostic metric for quantifying neuron-level learning capacity. We show that GraMa effectively reveals persistent neuron inactivity across diverse architectures, including residual networks, diffusion models, and agents with varied activation functions. Moreover, resetting neurons guided by GraMa (ReGraMa) consistently improves learning performance across multiple deep RL algorithms and benchmarks, such as MuJoCo and the DeepMind Control Suite.

View on arXiv
@article{liu2025_2505.24061,
  title={ Measure gradients, not activations! Enhancing neuronal activity in deep reinforcement learning },
  author={ Jiashun Liu and Zihao Wu and Johan Obando-Ceron and Pablo Samuel Castro and Aaron Courville and Ling Pan },
  journal={arXiv preprint arXiv:2505.24061},
  year={ 2025 }
}
Comments on this paper