ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2502.19339
48
0

Evaluating LLMs and Pre-trained Models for Text Summarization Across Diverse Datasets

26 February 2025
Tohida Rehman
Soumabha Ghosh
Kuntal Das
Souvik Bhattacharjee
Debarshi Kumar Sanyal
S. Chattopadhyay
ArXivPDFHTML
Abstract

Text summarization plays a crucial role in natural language processing by condensing large volumes of text into concise and coherent summaries. As digital content continues to grow rapidly and the demand for effective information retrieval increases, text summarization has become a focal point of research in recent years. This study offers a thorough evaluation of four leading pre-trained and open-source large language models: BART, FLAN-T5, LLaMA-3-8B, and Gemma-7B, across five diverse datasets CNN/DM, Gigaword, News Summary, XSum, and BBC News. The evaluation employs widely recognized automatic metrics, including ROUGE-1, ROUGE-2, ROUGE-L, BERTScore, and METEOR, to assess the models' capabilities in generating coherent and informative summaries. The results reveal the comparative strengths and limitations of these models in processing various text types.

View on arXiv
@article{rehman2025_2502.19339,
  title={ Evaluating LLMs and Pre-trained Models for Text Summarization Across Diverse Datasets },
  author={ Tohida Rehman and Soumabha Ghosh and Kuntal Das and Souvik Bhattacharjee and Debarshi Kumar Sanyal and Samiran Chattopadhyay },
  journal={arXiv preprint arXiv:2502.19339},
  year={ 2025 }
}
Comments on this paper