72
0

Runaway is Ashamed, But Helpful: On the Early-Exit Behavior of Large Language Model-based Agents in Embodied Environments

Abstract

Agents powered by large language models (LLMs) have demonstrated strong planning and decision-making capabilities in complex embodied environments. However, such agents often suffer from inefficiencies in multi-turn interactions, frequently trapped in repetitive loops or issuing ineffective commands, leading to redundant computational overhead. Instead of relying solely on learning from trajectories, we take a first step toward exploring the early-exit behavior for LLM-based agents. We propose two complementary approaches: 1. an intrinsic\textbf{intrinsic} method that injects exit instructions during generation, and 2. an extrinsic\textbf{extrinsic} method that verifies task completion to determine when to halt an agent's trial. To evaluate early-exit mechanisms, we introduce two metrics: one measures the reduction of redundant steps\textbf{redundant steps} as a positive effect, and the other evaluates progress degradation\textbf{progress degradation} as a negative effect. Experiments with 4 different LLMs across 5 embodied environments show significant efficiency improvements, with only minor drops in agent performance. We also validate a practical strategy where a stronger agent assists after an early-exit agent, achieving better performance with the same total steps. We will release our code to support further research.

View on arXiv
@article{lu2025_2505.17616,
  title={ Runaway is Ashamed, But Helpful: On the Early-Exit Behavior of Large Language Model-based Agents in Embodied Environments },
  author={ Qingyu Lu and Liang Ding and Siyi Cao and Xuebo Liu and Kanjian Zhang and Jinxia Zhang and Dacheng Tao },
  journal={arXiv preprint arXiv:2505.17616},
  year={ 2025 }
}
Comments on this paper