ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2305.14251
  4. Cited By
FActScore: Fine-grained Atomic Evaluation of Factual Precision in Long
  Form Text Generation
v1v2 (latest)

FActScore: Fine-grained Atomic Evaluation of Factual Precision in Long Form Text Generation

Conference on Empirical Methods in Natural Language Processing (EMNLP), 2023
23 May 2023
Sewon Min
Kalpesh Krishna
Xinxi Lyu
M. Lewis
Anuj Kumar
Pang Wei Koh
Mohit Iyyer
Luke Zettlemoyer
Hannaneh Hajishirzi
    HILMALM
ArXiv (abs)PDFHTMLHuggingFace (2 upvotes)

Papers citing "FActScore: Fine-grained Atomic Evaluation of Factual Precision in Long Form Text Generation"

13 / 613 papers shown
Title
Siren's Song in the AI Ocean: A Survey on Hallucination in Large Language Models
Siren's Song in the AI Ocean: A Survey on Hallucination in Large Language ModelsComputational Linguistics (CL), 2023
Yue Zhang
Yafu Li
Leyang Cui
Deng Cai
Lemao Liu
...
Longyue Wang
Anh Tuan Luu
Freda Shi
Shuming Shi
Shuming Shi
LRMRALMHILM
646
790
0
03 Sep 2023
Halo: Estimation and Reduction of Hallucinations in Open-Source Weak
  Large Language Models
Halo: Estimation and Reduction of Hallucinations in Open-Source Weak Large Language Models
Mohamed S. Elaraby
Mengyin Lu
Jacob Dunn
Xueying Zhang
Yu Wang
Shizhu Liu
Pingchuan Tian
Yuping Wang
Yuxuan Wang
HILM
285
41
0
22 Aug 2023
Answering Unseen Questions With Smaller Language Models Using Rationale
  Generation and Dense Retrieval
Answering Unseen Questions With Smaller Language Models Using Rationale Generation and Dense Retrieval
Tim Hartill
Diana Benavides-Prado
Michael Witbrock
Patricia J. Riddle
ReLMLRM
156
3
0
09 Aug 2023
Automatically Correcting Large Language Models: Surveying the landscape
  of diverse self-correction strategies
Automatically Correcting Large Language Models: Surveying the landscape of diverse self-correction strategies
Liangming Pan
Michael Stephen Saxon
Wenda Xu
Deepak Nathani
Xinyi Wang
William Yang Wang
KELMLRM
355
262
0
06 Aug 2023
On the Trustworthiness Landscape of State-of-the-art Generative Models:
  A Survey and Outlook
On the Trustworthiness Landscape of State-of-the-art Generative Models: A Survey and OutlookInternational Journal of Computer Vision (IJCV), 2023
Mingyuan Fan
Chengyu Wang
Cen Chen
Yang Liu
Jun Huang
HILM
247
11
0
31 Jul 2023
FLASK: Fine-grained Language Model Evaluation based on Alignment Skill
  Sets
FLASK: Fine-grained Language Model Evaluation based on Alignment Skill SetsInternational Conference on Learning Representations (ICLR), 2023
Seonghyeon Ye
Doyoung Kim
Sungdong Kim
Hyeonbin Hwang
Seungone Kim
Yongrae Jo
James Thorne
Juho Kim
Minjoon Seo
ALM
527
143
0
20 Jul 2023
Generating Benchmarks for Factuality Evaluation of Language Models
Generating Benchmarks for Factuality Evaluation of Language ModelsConference of the European Chapter of the Association for Computational Linguistics (EACL), 2023
Dor Muhlgay
Ori Ram
Inbal Magar
Yoav Levine
Nir Ratner
Yonatan Belinkov
Omri Abend
Kevin Leyton-Brown
Amnon Shashua
Y. Shoham
HILM
183
122
0
13 Jul 2023
A Survey on Evaluation of Large Language Models
A Survey on Evaluation of Large Language ModelsACM Transactions on Intelligent Systems and Technology (ACM TIST), 2023
Yu-Chu Chang
Xu Wang
Yongfeng Zhang
Yuanyi Wu
Linyi Yang
...
Yue Zhang
Yi-Ju Chang
Philip S. Yu
Qian Yang
Xingxu Xie
ELMLM&MAALM
700
2,663
0
06 Jul 2023
Fine-Grained Human Feedback Gives Better Rewards for Language Model
  Training
Fine-Grained Human Feedback Gives Better Rewards for Language Model TrainingNeural Information Processing Systems (NeurIPS), 2023
Zeqiu Wu
Yushi Hu
Weijia Shi
Nouha Dziri
Alane Suhr
Prithviraj Ammanabrolu
Noah A. Smith
Mari Ostendorf
Hannaneh Hajishirzi
ALM
441
408
0
02 Jun 2023
Self-contradictory Hallucinations of Large Language Models: Evaluation,
  Detection and Mitigation
Self-contradictory Hallucinations of Large Language Models: Evaluation, Detection and MitigationInternational Conference on Learning Representations (ICLR), 2023
Niels Mündler
Jingxuan He
Slobodan Jenko
Martin Vechev
HILM
233
154
0
25 May 2023
Self-Checker: Plug-and-Play Modules for Fact-Checking with Large
  Language Models
Self-Checker: Plug-and-Play Modules for Fact-Checking with Large Language Models
Miaoran Li
Baolin Peng
Michel Galley
Jianfeng Gao
Zhu Zhang
LRMHILMKELM
203
38
0
24 May 2023
WikiChat: Stopping the Hallucination of Large Language Model Chatbots by
  Few-Shot Grounding on Wikipedia
WikiChat: Stopping the Hallucination of Large Language Model Chatbots by Few-Shot Grounding on WikipediaConference on Empirical Methods in Natural Language Processing (EMNLP), 2023
Sina J. Semnani
Violet Z. Yao
He Zhang
M. Lam
KELMAI4MH
283
101
0
23 May 2023
SelfCheckGPT: Zero-Resource Black-Box Hallucination Detection for
  Generative Large Language Models
SelfCheckGPT: Zero-Resource Black-Box Hallucination Detection for Generative Large Language ModelsConference on Empirical Methods in Natural Language Processing (EMNLP), 2023
Potsawee Manakul
Adian Liusie
Mark Gales
HILMLRM
409
639
0
15 Mar 2023
Previous
123...111213