182
v1v2 (latest)

ZeroDL: Zero-shot Distribution Learning for Text Clustering via Large Language Models

Annual Meeting of the Association for Computational Linguistics (ACL), 2024
Main:5 Pages
2 Figures
Bibliography:2 Pages
9 Tables
Appendix:4 Pages
Abstract

The advancements in large language models (LLMs) have brought significant progress in NLP tasks. However, if a task cannot be fully described in prompts, the models could fail to carry out the task. In this paper, we propose a simple yet effective method to contextualize a task toward a LLM. The method utilizes (1) open-ended zero-shot inference from the entire dataset, (2) aggregate the inference results, and (3) finally incorporate the aggregated meta-information for the actual task. We show the effectiveness in text clustering tasks, empowering LLMs to perform text-to-text-based clustering and leading to improvements on several datasets. Furthermore, we explore the generated class labels for clustering, showing how the LLM understands the task through data.

View on arXiv
Comments on this paper