ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2503.24340
79
1

Faster Rates for No-Regret Learning in General Games via Cautious Optimism

31 March 2025
Ashkan Soleymani
Georgios Piliouras
Gabriele Farina
ArXivPDFHTML
Abstract

We establish the first uncoupled learning algorithm that attains O(nlog⁡2dlog⁡T)O(n \log^2 d \log T)O(nlog2dlogT) per-player regret in multi-player general-sum games, where nnn is the number of players, ddd is the number of actions available to each player, and TTT is the number of repetitions of the game. Our results exponentially improve the dependence on ddd compared to the O(n dlog⁡T)O(n\, d \log T)O(ndlogT) regret attainable by Log-Regularized Lifted Optimistic FTRL [Far+22c], and also reduce the dependence on the number of iterations TTT from log⁡4T\log^4 Tlog4T to log⁡T\log TlogT compared to Optimistic Hedge, the previously well-studied algorithm with O(nlog⁡dlog⁡4T)O(n \log d \log^4 T)O(nlogdlog4T) regret [DFG21]. Our algorithm is obtained by combining the classic Optimistic Multiplicative Weights Update (OMWU) with an adaptive, non-monotonic learning rate that paces the learning process of the players, making them more cautious when their regret becomes too negative.

View on arXiv
@article{soleymani2025_2503.24340,
  title={ Faster Rates for No-Regret Learning in General Games via Cautious Optimism },
  author={ Ashkan Soleymani and Georgios Piliouras and Gabriele Farina },
  journal={arXiv preprint arXiv:2503.24340},
  year={ 2025 }
}
Comments on this paper