ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2505.01979
15
0

D3HRL: A Distributed Hierarchical Reinforcement Learning Approach Based on Causal Discovery and Spurious Correlation Detection

4 May 2025
Chenran Zhao
Dianxi Shi
Mengzhu Wang
Jianqiang Xia
Huanhuan Yang
Songchang Jin
Shaowu Yang
Chunping Qiu
ArXivPDFHTML
Abstract

Current Hierarchical Reinforcement Learning (HRL) algorithms excel in long-horizon sequential decision-making tasks but still face two challenges: delay effects and spurious correlations. To address them, we propose a causal HRL approach called D3HRL. First, D3HRL models delayed effects as causal relationships across different time spans and employs distributed causal discovery to learn these relationships. Second, it employs conditional independence testing to eliminate spurious correlations. Finally, D3HRL constructs and trains hierarchical policies based on the identified true causal relationships. These three steps are iteratively executed, gradually exploring the complete causal chain of the task. Experiments conducted in 2D-MineCraft and MiniGrid show that D3HRL demonstrates superior sensitivity to delay effects and accurately identifies causal relationships, leading to reliable decision-making in complex environments.

View on arXiv
@article{zhao2025_2505.01979,
  title={ D3HRL: A Distributed Hierarchical Reinforcement Learning Approach Based on Causal Discovery and Spurious Correlation Detection },
  author={ Chenran Zhao and Dianxi Shi and Mengzhu Wang and Jianqiang Xia and Huanhuan Yang and Songchang Jin and Shaowu Yang and Chunping Qiu },
  journal={arXiv preprint arXiv:2505.01979},
  year={ 2025 }
}
Comments on this paper