ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2505.00019
37
0

An Empirical Study on Prompt Compression for Large Language Models

24 April 2025
Z. Zhang
Jinyi Li
Yihuai Lan
X. Wang
Hao Wang
    MQ
ArXivPDFHTML
Abstract

Prompt engineering enables Large Language Models (LLMs) to perform a variety of tasks. However, lengthy prompts significantly increase computational complexity and economic costs. To address this issue, we study six prompt compression methods for LLMs, aiming to reduce prompt length while maintaining LLM response quality. In this paper, we present a comprehensive analysis covering aspects such as generation performance, model hallucinations, efficacy in multimodal tasks, word omission analysis, and more. We evaluate these methods across 13 datasets, including news, scientific articles, commonsense QA, math QA, long-context QA, and VQA datasets. Our experiments reveal that prompt compression has a greater impact on LLM performance in long contexts compared to short ones. In the Longbench evaluation, moderate compression even enhances LLM performance. Our code and data is available atthis https URL.

View on arXiv
@article{zhang2025_2505.00019,
  title={ An Empirical Study on Prompt Compression for Large Language Models },
  author={ Zheng Zhang and Jinyi Li and Yihuai Lan and Xiang Wang and Hao Wang },
  journal={arXiv preprint arXiv:2505.00019},
  year={ 2025 }
}
Comments on this paper