ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2511.01016
101
0
v1v2 (latest)

Prompt-R1: Collaborative Automatic Prompting Framework via End-to-end Reinforcement Learning

2 November 2025
Wenjin Liu
Haoran Luo
X. Lin
Haoming Liu
Tiesunlong Shen
Jiapu Wang
Rui Mao
Erik Cambria
    LLMAGOffRLLRM
ArXiv (abs)PDFHTMLGithub (29★)
Main:8 Pages
9 Figures
Bibliography:3 Pages
7 Tables
Appendix:10 Pages
Abstract

Recently, advanced large language models (LLMs) have emerged at an increasingly rapid pace. However, when faced with complex problems, most users are often unable to provide accurate and effective prompts to interact with LLMs, thus limiting the performance of LLMs. To address this challenge, we propose Prompt-R1, an end-to-end reinforcement learning framework that uses a small-scale LLM to collaborate with large-scale LLMs, replacing user interaction to solve problems better. This collaboration is cast as a multi-turn prompt interaction, where the small-scale LLM thinks and generates prompts, and the large-scale LLM performs complex reasoning. A dual-constrained reward is designed to optimize for correctness, generation quality, and reasoning accuracy. Prompt-R1 provides a plug-and-play framework that supports both inference and training with various large-scale LLMs. Experiments on multiple public datasets show that Prompt-R1 significantly outperforms baseline models across tasks. Our code is publicly available atthis https URL.

View on arXiv
Comments on this paper