ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2504.13500
27
0

Prejudge-Before-Think: Enhancing Large Language Models at Test-Time by Process Prejudge Reasoning

18 April 2025
J. T. Wang
Jin Jiang
Yang Liu
M. Zhang
Xunliang Cai
    LRM
ArXivPDFHTML
Abstract

In this paper, we introduce a new \emph{process prejudge} strategy in LLM reasoning to demonstrate that bootstrapping with process prejudge allows the LLM to adaptively anticipate the errors encountered when advancing the subsequent reasoning steps, similar to people sometimes pausing to think about what mistakes may occur and how to avoid them, rather than relying solely on trial and error. Specifically, we define a prejudge node in the rationale, which represents a reasoning step, with at least one step that follows the prejudge node that has no paths toward the correct answer. To synthesize the prejudge reasoning process, we present an automated reasoning framework with a dynamic tree-searching strategy. This framework requires only one LLM to perform answer judging, response critiquing, prejudge generation, and thought completion. Furthermore, we develop a two-phase training mechanism with supervised fine-tuning (SFT) and reinforcement learning (RL) to further enhance the reasoning capabilities of LLMs. Experimental results from competition-level complex reasoning demonstrate that our method can teach the model to prejudge before thinking and significantly enhance the reasoning ability of LLMs. Code and data is released atthis https URL.

View on arXiv
@article{wang2025_2504.13500,
  title={ Prejudge-Before-Think: Enhancing Large Language Models at Test-Time by Process Prejudge Reasoning },
  author={ Jianing Wang and Jin Jiang and Yang Liu and Mengdi Zhang and Xunliang Cai },
  journal={arXiv preprint arXiv:2504.13500},
  year={ 2025 }
}
Comments on this paper