ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2506.11083
  4. Cited By
RedDebate: Safer Responses through Multi-Agent Red Teaming Debates
v1v2 (latest)

RedDebate: Safer Responses through Multi-Agent Red Teaming Debates

4 June 2025
Ali Asad
Stephen Obadinma
Radin Shayanfar
Xiaodan Zhu
    AAMLLLMAG
ArXiv (abs)PDFHTMLGithub (1★)

Papers citing "RedDebate: Safer Responses through Multi-Agent Red Teaming Debates"

2 / 2 papers shown
Title
The Alignment Waltz: Jointly Training Agents to Collaborate for Safety
The Alignment Waltz: Jointly Training Agents to Collaborate for Safety
Jingyu Zhang
Haozhu Wang
Eric Michael Smith
Sid Wang
Amr Sharaf
Mahesh Pasupuleti
Benjamin Van Durme
Daniel Khashabi
Jason Weston
Hongyuan Zhan
88
0
0
09 Oct 2025
SafeEvalAgent: Toward Agentic and Self-Evolving Safety Evaluation of LLMs
SafeEvalAgent: Toward Agentic and Self-Evolving Safety Evaluation of LLMs
Yixu Wang
Xin Wang
Yang Yao
X. Li
Yan Teng
Jiabo He
Y. Wang
LLMAGELM
148
1
0
30 Sep 2025
1