ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2503.17933
76
0
v1v2 (latest)

Experience Retrieval-Augmentation with Electronic Health Records Enables Accurate Discharge QA

23 March 2025
Justice Ou
Tinglin Huang
Yilun Zhao
Ziyang Yu
Peiqing Lu
Rex Ying
    RALM
ArXiv (abs)PDFHTML
Main:8 Pages
8 Figures
Bibliography:2 Pages
8 Tables
Appendix:8 Pages
Abstract

To improve the reliability of Large Language Models (LLMs) in clinical applications, retrieval-augmented generation (RAG) is extensively applied to provide factual medical knowledge. However, beyond general medical knowledge from open-ended datasets, clinical case-based knowledge is also critical for effective medical reasoning, as it provides context grounded in real-world patient experiences. Motivated by this, we propose Experience Retrieval Augmentation - ExpRAG framework based on Electronic Health Record (EHR), aiming to offer the relevant context from other patients' discharge reports. ExpRAG performs retrieval through a coarse-to-fine process, utilizing an EHR-based report ranker to efficiently identify similar patients, followed by an experience retriever to extract task-relevant content for enhanced medical reasoning. To evaluate ExpRAG, we introduce DischargeQA, a clinical QA dataset with 1,280 discharge-related questions across diagnosis, medication, and instruction tasks. Each problem is generated using EHR data to ensure realistic and challenging scenarios. Experimental results demonstrate that ExpRAG consistently outperforms a text-based ranker, achieving an average relative improvement of 5.2%, highlighting the importance of case-based knowledge for medical reasoning.

View on arXiv
@article{ou2025_2503.17933,
  title={ Experience Retrieval-Augmentation with Electronic Health Records Enables Accurate Discharge QA },
  author={ Justice Ou and Tinglin Huang and Yilun Zhao and Ziyang Yu and Peiqing Lu and Rex Ying },
  journal={arXiv preprint arXiv:2503.17933},
  year={ 2025 }
}
Comments on this paper