ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2503.07237
50
0

LLM-C3MOD: A Human-LLM Collaborative System for Cross-Cultural Hate Speech Moderation

10 March 2025
Junyeong Park
Seogyeong Jeong
S.
Yohan Lee
Alice H. Oh
ArXivPDFHTML
Abstract

Content moderation is a global challenge, yet major tech platforms prioritize high-resource languages, leaving low-resource languages with scarce native moderators. Since effective moderation depends on understanding contextual cues, this imbalance increases the risk of improper moderation due to non-native moderators' limited cultural understanding. Through a user study, we identify that non-native moderators struggle with interpreting culturally-specific knowledge, sentiment, and internet culture in the hate speech moderation. To assist them, we present LLM-C3MOD, a human-LLM collaborative pipeline with three steps: (1) RAG-enhanced cultural context annotations; (2) initial LLM-based moderation; and (3) targeted human moderation for cases lacking LLM consensus. Evaluated on a Korean hate speech dataset with Indonesian and German participants, our system achieves 78% accuracy (surpassing GPT-4o's 71% baseline), while reducing human workload by 83.6%. Notably, human moderators excel at nuanced contents where LLMs struggle. Our findings suggest that non-native moderators, when properly supported by LLMs, can effectively contribute to cross-cultural hate speech moderation.

View on arXiv
@article{park2025_2503.07237,
  title={ LLM-C3MOD: A Human-LLM Collaborative System for Cross-Cultural Hate Speech Moderation },
  author={ Junyeong Park and Seogyeong Jeong and Seyoung Song and Yohan Lee and Alice Oh },
  journal={arXiv preprint arXiv:2503.07237},
  year={ 2025 }
}
Comments on this paper