DomainSum: A Hierarchical Benchmark for Fine-Grained Domain Shift in
Abstractive Text Summarization
North American Chapter of the Association for Computational Linguistics (NAACL), 2024
Main:8 Pages
5 Figures
Bibliography:3 Pages
8 Tables
Appendix:2 Pages
Abstract
Most research on abstractive summarization focuses on single-domain applications, often neglecting how domain shifts between documents affect performance and the generalization ability of summarization models. To address this issue, we introduce DomainSum, a hierarchical benchmark designed to capture fine-grained domain shifts in abstractive summarization. We categorize these shifts into three levels: genre, style, and topic, and demonstrate through comprehensive benchmark analysis that they follow a hierarchical structure. Furthermore, we evaluate the domain generalization capabilities of commonly used pre-trained language models (PLMs) and large language models (LLMs) in in-domain and cross-domain settings.
View on arXivComments on this paper
