ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2503.10304
36
0

Nash Equilibrium Constrained Auto-bidding With Bi-level Reinforcement Learning

13 March 2025
Zhiyu Mou
Miao Xu
Rongquan Bai
Zhuoran Yang
Chuan Yu
Jian Xu
Bo Zheng
ArXivPDFHTML
Abstract

Many online advertising platforms provide advertisers with auto-bidding services to enhance their advertising performance. However, most existing auto-bidding algorithms fail to accurately capture the auto-bidding problem formulation that the platform truly faces, let alone solve it. Actually, we argue that the platform should try to help optimize each advertiser's performance to the greatest extent -- which makes ϵ\epsilonϵ-Nash Equilibrium (ϵ\epsilonϵ-NE) a necessary solution concept -- while maximizing the social welfare of all the advertisers for the platform's long-term value. Based on this, we introduce the \emph{Nash-Equilibrium Constrained Bidding} (NCB), a new formulation of the auto-bidding problem from the platform's perspective. Specifically, it aims to maximize the social welfare of all advertisers under the ϵ\epsilonϵ-NE constraint. However, the NCB problem presents significant challenges due to its constrained bi-level structure and the typically large number of advertisers involved. To address these challenges, we propose a \emph{Bi-level Policy Gradient} (BPG) framework with theoretical guarantees. Notably, its computational complexity is independent of the number of advertisers, and the associated gradients are straightforward to compute. Extensive simulated and real-world experiments validate the effectiveness of the BPG framework.

View on arXiv
@article{mou2025_2503.10304,
  title={ Nash Equilibrium Constrained Auto-bidding With Bi-level Reinforcement Learning },
  author={ Zhiyu Mou and Miao Xu and Rongquan Bai and Zhuoran Yang and Chuan Yu and Jian Xu and Bo Zheng },
  journal={arXiv preprint arXiv:2503.10304},
  year={ 2025 }
}
Comments on this paper