Policy Mirror Ascent for Efficient and Independent Learning in Mean
Field Games
Mean-field games have been used as a theoretical tool to obtain an approximate Nash equilibrium for symmetric and anonymous -player games in literature. However, limiting applicability, existing theoretical results assume variations of a "population generative model", which allows arbitrary modifications of the population distribution by the learning algorithm. Instead, we show that agents running policy mirror ascent converge to the Nash equilibrium of the regularized game within samples from a single sample trajectory without a population generative model, up to a standard error due to the mean field. Taking a divergent approach from literature, instead of working with the best-response map we first show that a policy mirror ascent map can be used to construct a contractive operator having the Nash equilibrium as its fixed point. Next, we prove that conditional TD-learning in -agent games can learn value functions within time steps. These results allow proving sample complexity guarantees in the oracle-free setting by only relying on a sample path from the agent simulator. Furthermore, we demonstrate that our methodology allows for independent learning by agents with finite sample guarantees.
View on arXiv