33
15

Long-horizon Locomotion and Manipulation on a Quadrupedal Robot with Large Language Models

Abstract

We present a large language model (LLM) based system to empower quadrupedal robots with problem-solving abilities for long-horizon tasks beyond short-term motions. Long-horizon tasks for quadrupeds are challenging since they require both a high-level understanding of the semantics of the problem for task planning and a broad range of locomotion and manipulation skills to interact with the environment. Our system builds a high-level reasoning layer with large language models, which generates hybrid discrete-continuous plans as robot code from task descriptions. It comprises multiple LLM agents: a semantic planner that sketches a plan, a parameter calculator that predicts arguments in the plan, a code generator that converts the plan into executable robot code, and a replanner that handles execution failures or human interventions. At the low level, we adopt reinforcement learning to train a set of motion planning and control skills to unleash the flexibility of quadrupeds for rich environment interactions. Our system is tested on long-horizon tasks that are infeasible to complete with one single skill. Simulation and real-world experiments show that it successfully figures out multi-step strategies and demonstrates non-trivial behaviors, including building tools or notifying a human for help. Demos are available on our project page:this https URL.

View on arXiv
@article{ouyang2025_2404.05291,
  title={ Long-horizon Locomotion and Manipulation on a Quadrupedal Robot with Large Language Models },
  author={ Yutao Ouyang and Jinhan Li and Yunfei Li and Zhongyu Li and Chao Yu and Koushil Sreenath and Yi Wu },
  journal={arXiv preprint arXiv:2404.05291},
  year={ 2025 }
}
Comments on this paper