ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2504.04191
33
1

GROVE: A Generalized Reward for Learning Open-Vocabulary Physical Skill

5 April 2025
Jieming Cui
Tengyu Liu
Ziyu Meng
Jiale Yu
Ran Song
Wei Zhang
Yixin Zhu
Siyuan Huang
    VLM
ArXivPDFHTML
Abstract

Learning open-vocabulary physical skills for simulated agents presents a significant challenge in artificial intelligence. Current reinforcement learning approaches face critical limitations: manually designed rewards lack scalability across diverse tasks, while demonstration-based methods struggle to generalize beyond their training distribution. We introduce GROVE, a generalized reward framework that enables open-vocabulary physical skill learning without manual engineering or task-specific demonstrations. Our key insight is that Large Language Models(LLMs) and Vision Language Models(VLMs) provide complementary guidance -- LLMs generate precise physical constraints capturing task requirements, while VLMs evaluate motion semantics and naturalness. Through an iterative design process, VLM-based feedback continuously refines LLM-generated constraints, creating a self-improving reward system. To bridge the domain gap between simulation and natural images, we develop Pose2CLIP, a lightweight mapper that efficiently projects agent poses directly into semantic feature space without computationally expensive rendering. Extensive experiments across diverse embodiments and learning paradigms demonstrate GROVE's effectiveness, achieving 22.2% higher motion naturalness and 25.7% better task completion scores while training 8.4x faster than previous methods. These results establish a new foundation for scalable physical skill acquisition in simulated environments.

View on arXiv
@article{cui2025_2504.04191,
  title={ GROVE: A Generalized Reward for Learning Open-Vocabulary Physical Skill },
  author={ Jieming Cui and Tengyu Liu and Ziyu Meng and Jiale Yu and Ran Song and Wei Zhang and Yixin Zhu and Siyuan Huang },
  journal={arXiv preprint arXiv:2504.04191},
  year={ 2025 }
}
Comments on this paper