Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2310.09624
Cited By
ASSERT: Automated Safety Scenario Red Teaming for Evaluating the Robustness of Large Language Models
14 October 2023
Alex Mei
Sharon Levy
William Yang Wang
AAML
Re-assign community
ArXiv
PDF
HTML
Papers citing
"ASSERT: Automated Safety Scenario Red Teaming for Evaluating the Robustness of Large Language Models"
3 / 3 papers shown
Title
Be a Multitude to Itself: A Prompt Evolution Framework for Red Teaming
Rui Li
Peiyi Wang
Jingyuan Ma
Di Zhang
Lei Sha
Zhifang Sui
LLMAG
44
0
0
22 Feb 2025
Training language models to follow instructions with human feedback
Long Ouyang
Jeff Wu
Xu Jiang
Diogo Almeida
Carroll L. Wainwright
...
Amanda Askell
Peter Welinder
Paul Christiano
Jan Leike
Ryan J. Lowe
OSLM
ALM
301
11,730
0
04 Mar 2022
Fantastically Ordered Prompts and Where to Find Them: Overcoming Few-Shot Prompt Order Sensitivity
Yao Lu
Max Bartolo
Alastair Moore
Sebastian Riedel
Pontus Stenetorp
AILaw
LRM
274
882
0
18 Apr 2021
1