ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2503.04202
50
0

Computational Intractability of Strategizing against Online Learners

6 March 2025
A. Assos
Y. Dagan
Nived Rajaraman
ArXivPDFHTML
Abstract

Online learning algorithms are widely used in strategic multi-agent settings, including repeated auctions, contract design, and pricing competitions, where agents adapt their strategies over time. A key question in such environments is how an optimizing agent can best respond to a learning agent to improve its own long-term outcomes. While prior work has developed efficient algorithms for the optimizer in special cases - such as structured auction settings or contract design - no general efficient algorithm is known.In this paper, we establish a strong computational hardness result: unless P=NP\mathsf{P} = \mathsf{NP}P=NP, no polynomial-time optimizer can compute a near-optimal strategy against a learner using a standard no-regret algorithm, specifically Multiplicative Weights Update (MWU). Our result proves an Ω(T)\Omega(T)Ω(T) hardness bound, significantly strengthening previous work that only showed an additive Θ(1)\Theta(1)Θ(1) impossibility result. Furthermore, while the prior hardness result focused on learners using fictitious play - an algorithm that is not no-regret - we prove intractability for a widely used no-regret learning algorithm. This establishes a fundamental computational barrier to finding optimal strategies in general game-theoretic settings.

View on arXiv
@article{assos2025_2503.04202,
  title={ Computational Intractability of Strategizing against Online Learners },
  author={ Angelos Assos and Yuval Dagan and Nived Rajaraman },
  journal={arXiv preprint arXiv:2503.04202},
  year={ 2025 }
}
Comments on this paper