ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2303.10430
  4. Cited By
NoisyHate: Mining Online Human-Written Perturbations for Realistic Robustness Benchmarking of Content Moderation Models

NoisyHate: Mining Online Human-Written Perturbations for Realistic Robustness Benchmarking of Content Moderation Models

18 March 2023
Yiran Ye
Thai Le
Dongwon Lee
    AAML
    DeLMO
ArXivPDFHTML

Papers citing "NoisyHate: Mining Online Human-Written Perturbations for Realistic Robustness Benchmarking of Content Moderation Models"

2 / 2 papers shown
Title
Efficient Toxic Content Detection by Bootstrapping and Distilling Large
  Language Models
Efficient Toxic Content Detection by Bootstrapping and Distilling Large Language Models
Jiang Zhang
Qiong Wu
Yiming Xu
Cheng Cao
Zheng Du
Konstantinos Psounis
28
14
0
13 Dec 2023
Generating Natural Language Adversarial Examples
Generating Natural Language Adversarial Examples
M. Alzantot
Yash Sharma
Ahmed Elgohary
Bo-Jhang Ho
Mani B. Srivastava
Kai-Wei Chang
AAML
245
914
0
21 Apr 2018
1