ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2405.14445
22
4

Exploring the use of a Large Language Model for data extraction in systematic reviews: a rapid feasibility study

23 May 2024
Lena Schmidt
Kaitlyn Hair
Sergio Graziozi
Fiona Campbell
Claudia Kapp
Alireza Khanteymoori
Dawn Craig
Mark Engelbert
James Thomas
ArXivPDFHTML
Abstract

This paper describes a rapid feasibility study of using GPT-4, a large language model (LLM), to (semi)automate data extraction in systematic reviews. Despite the recent surge of interest in LLMs there is still a lack of understanding of how to design LLM-based automation tools and how to robustly evaluate their performance. During the 2023 Evidence Synthesis Hackathon we conducted two feasibility studies. Firstly, to automatically extract study characteristics from human clinical, animal, and social science domain studies. We used two studies from each category for prompt-development; and ten for evaluation. Secondly, we used the LLM to predict Participants, Interventions, Controls and Outcomes (PICOs) labelled within 100 abstracts in the EBM-NLP dataset. Overall, results indicated an accuracy of around 80%, with some variability between domains (82% for human clinical, 80% for animal, and 72% for studies of human social sciences). Causal inference methods and study design were the data extraction items with the most errors. In the PICO study, participants and intervention/control showed high accuracy (>80%), outcomes were more challenging. Evaluation was done manually; scoring methods such as BLEU and ROUGE showed limited value. We observed variability in the LLMs predictions and changes in response quality. This paper presents a template for future evaluations of LLMs in the context of data extraction for systematic review automation. Our results show that there might be value in using LLMs, for example as second or third reviewers. However, caution is advised when integrating models such as GPT-4 into tools. Further research on stability and reliability in practical settings is warranted for each type of data that is processed by the LLM.

View on arXiv
@article{schmidt2025_2405.14445,
  title={ Exploring the use of a Large Language Model for data extraction in systematic reviews: a rapid feasibility study },
  author={ Lena Schmidt and Kaitlyn Hair and Sergio Graziosi and Fiona Campbell and Claudia Kapp and Alireza Khanteymoori and Dawn Craig and Mark Engelbert and James Thomas },
  journal={arXiv preprint arXiv:2405.14445},
  year={ 2025 }
}
Comments on this paper