ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2503.20968
31
0

Reinforcement Learning for Efficient Toxicity Detection in Competitive Online Video Games

26 March 2025
Jacob Morrier
Rafal Kocielnik
R. Alvarez
    OffRL
ArXivPDFHTML
Abstract

Online platforms take proactive measures to detect and address undesirable behavior, aiming to focus these resource-intensive efforts where such behavior is most prevalent. This article considers the problem of efficient sampling for toxicity detection in competitive online video games. To make optimal monitoring decisions, video game service operators need estimates of the likelihood of toxic behavior. If no model is available for these predictions, one must be estimated in real time. To close this gap, we propose a contextual bandit algorithm that makes monitoring decisions based on a small set of variables that, according to domain expertise, are associated with toxic behavior. This algorithm balances exploration and exploitation to optimize long-term outcomes and is deliberately designed for easy deployment in production. Using data from the popular first-person action game Call of Duty: Modern Warfare III, we show that our algorithm consistently outperforms baseline algorithms that rely solely on players' past behavior. This finding has substantive implications for the nature of toxicity. It also illustrates how domain expertise can be harnessed to help video game service operators identify and mitigate toxicity, ultimately fostering a safer and more enjoyable gaming experience.

View on arXiv
@article{morrier2025_2503.20968,
  title={ Reinforcement Learning for Efficient Toxicity Detection in Competitive Online Video Games },
  author={ Jacob Morrier and Rafal Kocielnik and R. Michael Alvarez },
  journal={arXiv preprint arXiv:2503.20968},
  year={ 2025 }
}
Comments on this paper