ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2204.05961
  4. Cited By
Quantified Reproducibility Assessment of NLP Results

Quantified Reproducibility Assessment of NLP Results

12 April 2022
Anya Belz
Maja Popović
Simon Mille
ArXivPDFHTML

Papers citing "Quantified Reproducibility Assessment of NLP Results"

9 / 9 papers shown
Title
Principles from Clinical Research for NLP Model Generalization
Principles from Clinical Research for NLP Model Generalization
Aparna Elangovan
Jiayuan He
Yuan Li
Karin Verspoor
CML
27
3
0
07 Nov 2023
Collect, Measure, Repeat: Reliability Factors for Responsible AI Data
  Collection
Collect, Measure, Repeat: Reliability Factors for Responsible AI Data Collection
Oana Inel
Tim Draws
Lora Aroyo
38
6
0
22 Aug 2023
Position: Key Claims in LLM Research Have a Long Tail of Footnotes
Position: Key Claims in LLM Research Have a Long Tail of Footnotes
Anna Rogers
A. Luccioni
45
19
0
14 Aug 2023
With a Little Help from the Authors: Reproducing Human Evaluation of an
  MT Error Detector
With a Little Help from the Authors: Reproducing Human Evaluation of an MT Error Detector
Ondvrej Plátek
Mateusz Lango
Ondrej Dusek
22
3
0
12 Aug 2023
KoLA: Carefully Benchmarking World Knowledge of Large Language Models
KoLA: Carefully Benchmarking World Knowledge of Large Language Models
Jifan Yu
Xiaozhi Wang
Shangqing Tu
S. Cao
Daniel Zhang-Li
...
Lei Hou
Zhiyuan Liu
Bin Xu
Jie Tang
Juanzi Li
ELM
ALM
36
66
0
15 Jun 2023
Investigating Reproducibility at Interspeech Conferences: A Longitudinal
  and Comparative Perspective
Investigating Reproducibility at Interspeech Conferences: A Longitudinal and Comparative Perspective
Mohammad Arvan
A. Seza Doğruöz
Natalie Parde
11
0
0
07 Jun 2023
Missing Information, Unresponsive Authors, Experimental Flaws: The
  Impossibility of Assessing the Reproducibility of Previous Human Evaluations
  in NLP
Missing Information, Unresponsive Authors, Experimental Flaws: The Impossibility of Assessing the Reproducibility of Previous Human Evaluations in NLP
Anya Belz
Craig Thomson
Ehud Reiter
Gavin Abercrombie
J. Alonso-Moral
...
Antonio Toral
Xiao-Yi Wan
Leo Wanner
Lewis J. Watson
Diyi Yang
66
35
0
02 May 2023
When Good and Reproducible Results are a Giant with Feet of Clay: The
  Importance of Software Quality in NLP
When Good and Reproducible Results are a Giant with Feet of Clay: The Importance of Software Quality in NLP
Sara Papi
Marco Gaido
Andrea Pilzer
Matteo Negri
45
10
0
28 Mar 2023
Reproducibility Issues for BERT-based Evaluation Metrics
Reproducibility Issues for BERT-based Evaluation Metrics
Yanran Chen
Jonas Belouadi
Steffen Eger
36
16
0
30 Mar 2022
1