342
v1v2v3v4 (latest)

PromptSuite: A Task-Agnostic Framework for Multi-Prompt Generation

Main:6 Pages
6 Figures
Bibliography:2 Pages
4 Tables
Appendix:2 Pages
Abstract

Evaluating LLMs with a single prompt has proven unreliable, with small changes leading to significant performance differences. However, generating the prompt variations needed for a more robust multi-prompt evaluation is challenging, limiting its adoption in practice. To address this, we introduce PromptSuite, a framework that enables the automatic generation of various prompts. PromptSuite is flexible - working out of the box on a wide range of tasks and benchmarks. It follows a modular prompt design, allowing controlled perturbations to each component, and is extensible, supporting the addition of new components and perturbation types. Through a series of case studies, we show that PromptSuite provides meaningful variations to support strong evaluation practices. All resources, including the Python API, source code, user-friendly web interface, and demonstration video, are available at:this https URL.

View on arXiv
Comments on this paper