ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2306.11247
  4. Cited By
DICES Dataset: Diversity in Conversational AI Evaluation for Safety

DICES Dataset: Diversity in Conversational AI Evaluation for Safety

20 June 2023
Lora Aroyo
Alex S. Taylor
Mark Díaz
Christopher Homan
Alicia Parrish
Greg Serapio-García
Vinodkumar Prabhakaran
Ding Wang
ArXivPDFHTML

Papers citing "DICES Dataset: Diversity in Conversational AI Evaluation for Safety"

9 / 9 papers shown
Title
From Reviews to Dialogues: Active Synthesis for Zero-Shot LLM-based Conversational Recommender System
From Reviews to Dialogues: Active Synthesis for Zero-Shot LLM-based Conversational Recommender System
Rohan Surana
Junda Wu
Zhouhang Xie
Yu Xia
Harald Steck
Dawen Liang
Nathan Kallus
Julian McAuley
26
0
0
21 Apr 2025
Values in the Wild: Discovering and Analyzing Values in Real-World Language Model Interactions
Values in the Wild: Discovering and Analyzing Values in Real-World Language Model Interactions
Saffron Huang
Esin Durmus
Miles McCain
Kunal Handa
Alex Tamkin
Jerry Hong
Michael Stern
Arushi Somani
Xiuruo Zhang
Deep Ganguli
VLM
42
1
0
21 Apr 2025
Beyond correlation: The Impact of Human Uncertainty in Measuring the Effectiveness of Automatic Evaluation and LLM-as-a-Judge
Beyond correlation: The Impact of Human Uncertainty in Measuring the Effectiveness of Automatic Evaluation and LLM-as-a-Judge
Aparna Elangovan
Jongwoo Ko
Lei Xu
Mahsa Elyasi
Ling Liu
S. Bodapati
Dan Roth
41
5
0
28 Jan 2025
SafetyPrompts: a Systematic Review of Open Datasets for Evaluating and Improving Large Language Model Safety
SafetyPrompts: a Systematic Review of Open Datasets for Evaluating and Improving Large Language Model Safety
Paul Röttger
Fabio Pernisi
Bertie Vidgen
Dirk Hovy
ELM
KELM
55
30
0
08 Apr 2024
How Far Can We Extract Diverse Perspectives from Large Language Models?
How Far Can We Extract Diverse Perspectives from Large Language Models?
Shirley Anugrah Hayati
Minhwa Lee
Dheeraj Rajagopal
Dongyeop Kang
38
10
0
16 Nov 2023
Mitigating Toxic Degeneration with Empathetic Data: Exploring the
  Relationship Between Toxicity and Empathy
Mitigating Toxic Degeneration with Empathetic Data: Exploring the Relationship Between Toxicity and Empathy
Allison Lahnala
Charles F Welch
Béla Neuendorf
Lucie Flek
51
13
0
15 May 2022
Unsolved Problems in ML Safety
Unsolved Problems in ML Safety
Dan Hendrycks
Nicholas Carlini
John Schulman
Jacob Steinhardt
173
272
0
28 Sep 2021
Agreeing to Disagree: Annotating Offensive Language Datasets with
  Annotators' Disagreement
Agreeing to Disagree: Annotating Offensive Language Datasets with Annotators' Disagreement
Elisa Leonardelli
Stefano Menini
Alessio Palmero Aprosio
Marco Guerini
Sara Tonelli
45
97
0
28 Sep 2021
Mitigating Racial Biases in Toxic Language Detection with an
  Equity-Based Ensemble Framework
Mitigating Racial Biases in Toxic Language Detection with an Equity-Based Ensemble Framework
Matan Halevy
Camille Harris
A. Bruckman
Diyi Yang
A. Howard
34
35
0
27 Sep 2021
1