CAIMAN: Causal Action Influence Detection for Sample-efficient Loco-manipulation

Enabling legged robots to perform non-prehensile loco-manipulation is crucial for enhancing their versatility. Learning behaviors such as whole-body object pushing often requires sophisticated planning strategies or extensive task-specific reward shaping, especially in unstructured environments. In this work, we present CAIMAN, a practical reinforcement learning framework that encourages the agent to gain control over other entities in the environment. CAIMAN leverages causal action influence as an intrinsic motivation objective, allowing legged robots to efficiently acquire object pushing skills even under sparse task rewards. We employ a hierarchical control strategy, combining a low-level locomotion module with a high-level policy that generates task-relevant velocity commands and is trained to maximize the intrinsic reward. To estimate causal action influence, we learn the dynamics of the environment by integrating a kinematic prior with data collected duringthis http URLempirically demonstrate CAIMAN's superior sample efficiency and adaptability to diverse scenarios in simulation, as well as its successful transfer to real-world systems without further fine-tuning.
View on arXiv@article{yuan2025_2502.00835, title={ CAIMAN: Causal Action Influence Detection for Sample-efficient Loco-manipulation }, author={ Yuanchen Yuan and Jin Cheng and Núria Armengol Urpí and Stelian Coros }, journal={arXiv preprint arXiv:2502.00835}, year={ 2025 } }