ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2402.15987
  4. Cited By
Likelihood-based Mitigation of Evaluation Bias in Large Language Models

Likelihood-based Mitigation of Evaluation Bias in Large Language Models

25 February 2024
Masanari Ohi
Masahiro Kaneko
Ryuto Koike
Mengsay Loem
Naoaki Okazaki
ArXivPDFHTML

Papers citing "Likelihood-based Mitigation of Evaluation Bias in Large Language Models"

7 / 7 papers shown
Title
Perplexity Trap: PLM-Based Retrievers Overrate Low Perplexity Documents
Haoyu Wang
Sunhao Dai
Haiyuan Zhao
Liang Pang
Xiao Zhang
Gang Wang
Zhenhua Dong
Jun Xu
Ji-Rong Wen
56
2
0
11 Mar 2025
What Makes a Good Story and How Can We Measure It? A Comprehensive
  Survey of Story Evaluation
What Makes a Good Story and How Can We Measure It? A Comprehensive Survey of Story Evaluation
Dingyi Yang
Qin Jin
36
5
0
26 Aug 2024
A Little Leak Will Sink a Great Ship: Survey of Transparency for Large
  Language Models from Start to Finish
A Little Leak Will Sink a Great Ship: Survey of Transparency for Large Language Models from Start to Finish
Masahiro Kaneko
Timothy Baldwin
PILM
23
3
0
24 Mar 2024
A Survey on Human-AI Teaming with Large Pre-Trained Models
A Survey on Human-AI Teaming with Large Pre-Trained Models
Vanshika Vats
Marzia Binta Nizam
Minghao Liu
Ziyuan Wang
Richard Ho
...
Celeste Shen
Rachel Shen
Nafisa Hussain
Kesav Ravichandran
James Davis
LM&MA
36
8
0
07 Mar 2024
Evaluating Gender Bias of Pre-trained Language Models in Natural
  Language Inference by Considering All Labels
Evaluating Gender Bias of Pre-trained Language Models in Natural Language Inference by Considering All Labels
Panatchakorn Anantaprayoon
Masahiro Kaneko
Naoaki Okazaki
56
16
0
18 Sep 2023
The Impact of Debiasing on the Performance of Language Models in
  Downstream Tasks is Underestimated
The Impact of Debiasing on the Performance of Language Models in Downstream Tasks is Underestimated
Masahiro Kaneko
Danushka Bollegala
Naoaki Okazaki
26
5
0
16 Sep 2023
Can Large Language Models Be an Alternative to Human Evaluations?
Can Large Language Models Be an Alternative to Human Evaluations?
Cheng-Han Chiang
Hung-yi Lee
ALM
LM&MA
209
568
0
03 May 2023
1