ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2302.05711
  4. Cited By
Fair Enough: Standardizing Evaluation and Model Selection for Fairness
  Research in NLP

Fair Enough: Standardizing Evaluation and Model Selection for Fairness Research in NLP

11 February 2023
Xudong Han
Timothy Baldwin
Trevor Cohn
ArXivPDFHTML

Papers citing "Fair Enough: Standardizing Evaluation and Model Selection for Fairness Research in NLP"

5 / 5 papers shown
Title
It's complicated. The relationship of algorithmic fairness and non-discrimination regulations in the EU AI Act
It's complicated. The relationship of algorithmic fairness and non-discrimination regulations in the EU AI Act
Kristof Meding
FaML
64
1
0
22 Jan 2025
Unlearning as multi-task optimization: A normalized gradient difference approach with an adaptive learning rate
Unlearning as multi-task optimization: A normalized gradient difference approach with an adaptive learning rate
Zhiqi Bu
Xiaomeng Jin
Bhanukiran Vinzamuri
Anil Ramakrishna
Kai-Wei Chang
V. Cevher
Mingyi Hong
MU
83
6
0
29 Oct 2024
CHiSafetyBench: A Chinese Hierarchical Safety Benchmark for Large
  Language Models
CHiSafetyBench: A Chinese Hierarchical Safety Benchmark for Large Language Models
Wenjing Zhang
Xuejiao Lei
Zhaoxiang Liu
Meijuan An
Bikun Yang
Kaikai Zhao
Kai Wang
Shiguo Lian
ELM
34
7
0
14 Jun 2024
Evaluating Debiasing Techniques for Intersectional Biases
Evaluating Debiasing Techniques for Intersectional Biases
Shivashankar Subramanian
Xudong Han
Timothy Baldwin
Trevor Cohn
Lea Frermann
77
49
0
21 Sep 2021
Fair prediction with disparate impact: A study of bias in recidivism
  prediction instruments
Fair prediction with disparate impact: A study of bias in recidivism prediction instruments
Alexandra Chouldechova
FaML
185
2,082
0
24 Oct 2016
1