ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2003.00433
58
24
v1v2v3 (latest)

Asynchronous Policy Evaluation in Distributed Reinforcement Learning over Networks

1 March 2020
Xingyu Sha
Jiaqi Zhang
Keyou You
Kai Zhang
Tamer Basar
    OffRL
ArXiv (abs)PDFHTML
Abstract

This paper proposes a \emph{fully asynchronous} scheme for policy evaluation of distributed reinforcement learning (DisRL) over peer-to-peer networks. Without any form of coordination, nodes can communicate with neighbors and compute their local variables using (possibly) delayed information at any time, which is in sharp contrast to the asynchronous gossip. Thus, the proposed scheme fully takes advantage of the distributed setting. We prove that our method converges at a linear rate O(ck)\mathcal{O}(c^k)O(ck) where c∈(0,1)c\in(0,1)c∈(0,1) and kkk increases by one no matter on which node updates, showing the computational advantage by reducing the amount of synchronization. Numerical experiments show that our method speeds up linearly w.r.t. the number of nodes, and is robust to straggler nodes. To the best of our knowledge, our work is the first theoretical analysis for asynchronous update in DisRL, including the \emph{parallel RL} domain advocated by A3C.

View on arXiv
Comments on this paper