Credit Assignment and Efficient Exploration based on Influence Scope in Multi-agent Reinforcement Learning

Training cooperative agents in sparse-reward scenarios poses significant challenges for multi-agent reinforcement learning (MARL). Without clear feedback on actions at each step in sparse-reward setting, previous methods struggle with precise credit assignment among agents and effective exploration. In this paper, we introduce a novel method to deal with both credit assignment and exploration problems in reward-sparse domains. Accordingly, we propose an algorithm that calculates the Influence Scope of Agents (ISA) on states by taking specific value of the dimensions/attributes of states that can be influenced by individual agents. The mutual dependence between agents' actions and state attributes are then used to calculate the credit assignment and to delimit the exploration space for each individual agent. We then evaluate ISA in a variety of sparse-reward multi-agent scenarios. The results show that our method significantly outperforms the state-of-art baselines.
View on arXiv@article{han2025_2505.08630, title={ Credit Assignment and Efficient Exploration based on Influence Scope in Multi-agent Reinforcement Learning }, author={ Shuai Han and Mehdi Dastani and Shihan Wang }, journal={arXiv preprint arXiv:2505.08630}, year={ 2025 } }