ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2212.12567
30
9

Adapting to game trees in zero-sum imperfect information games

23 December 2022
Côme Fiegel
Pierre Ménard
Tadashi Kozuno
Rémi Munos
Vianney Perchet
Michal Valko
ArXivPDFHTML
Abstract

Imperfect information games (IIG) are games in which each player only partially observes the current game state. We study how to learn ϵ\epsilonϵ-optimal strategies in a zero-sum IIG through self-play with trajectory feedback. We give a problem-independent lower bound O~(H(AX+BY)/ϵ2)\widetilde{\mathcal{O}}(H(A_{\mathcal{X}}+B_{\mathcal{Y}})/\epsilon^2)O(H(AX​+BY​)/ϵ2) on the required number of realizations to learn these strategies with high probability, where HHH is the length of the game, AXA_{\mathcal{X}}AX​ and BYB_{\mathcal{Y}}BY​ are the total number of actions for the two players. We also propose two Follow the Regularized leader (FTRL) algorithms for this setting: Balanced FTRL which matches this lower bound, but requires the knowledge of the information set structure beforehand to define the regularization; and Adaptive FTRL which needs O~(H2(AX+BY)/ϵ2)\widetilde{\mathcal{O}}(H^2(A_{\mathcal{X}}+B_{\mathcal{Y}})/\epsilon^2)O(H2(AX​+BY​)/ϵ2) realizations without this requirement by progressively adapting the regularization to the observations.

View on arXiv
Comments on this paper