v1v2 (latest)
A finite time analysis of distributed Q-learning
- OffRL
Main:12 Pages
3 Figures
Bibliography:5 Pages
1 Tables
Appendix:19 Pages
Abstract
Multi-agent reinforcement learning (MARL) has witnessed a remarkable surge in interest, fueled by the empirical success achieved in applications of single-agent reinforcement learning (RL). In this study, we consider a distributed Q-learning scenario, wherein a number of agents cooperatively solve a sequential decision making problem without access to the central reward function which is an average of the local rewards. In particular, we study finite-time analysis of a distributed Q-learning algorithm, and provide a new sample complexity result of under tabular lookup
View on arXivComments on this paper
