587
v1v2v3 (latest)

Verbosity-Aware Rationale Reduction: Effective Reduction of Redundant Rationale via Principled Criteria

Annual Meeting of the Association for Computational Linguistics (ACL), 2024
Main:9 Pages
7 Figures
Bibliography:3 Pages
11 Tables
Appendix:4 Pages
Abstract

Large Language Models (LLMs) rely on generating extensive intermediate reasoning units (e.g., tokens, sentences) to enhance final answer quality across a wide range of complex tasks. While this approach has proven effective, it inevitably increases substantial inference costs. Previous methods adopting token-level reduction without clear criteria result in poor performance compared to models trained with complete rationale. To address this challenge, we propose a novel sentence-level rationale reduction framework leveraging likelihood-based criteria, verbosity, to identify and remove redundant reasoning sentences. Unlike previous approaches, our method leverages verbosity to selectively remove redundant reasoning sentences while preserving reasoning capabilities. Our experimental results across various reasoning tasks demonstrate that our method improves performance by an average of 7.71% while reducing token generation by 19.87% compared to model trained with complete reasoning paths.

View on arXiv
Comments on this paper