ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2503.08007
47
1

MoRE: Unlocking Scalability in Reinforcement Learning for Quadruped Vision-Language-Action Models

11 March 2025
H. Zhao
Wenxuan Song
Donglin Wang
Xinyang Tong
Pengxiang Ding
Xuelian Cheng
Zongyuan Ge
ArXivPDFHTML
Abstract

Developing versatile quadruped robots that can smoothly perform various actions and tasks in real-world environments remains a significant challenge. This paper introduces a novel vision-language-action (VLA) model, mixture of robotic experts (MoRE), for quadruped robots that aim to introduce reinforcement learning (RL) for fine-tuning large-scale VLA models with a large amount of mixed-quality data. MoRE integrates multiple low-rank adaptation modules as distinct experts within a dense multi-modal large language model (MLLM), forming a sparse-activated mixture-of-experts model. This design enables the model to effectively adapt to a wide array of downstream tasks. Moreover, we employ a reinforcement learning-based training objective to train our model as a Q-function after deeply exploring the structural properties of our tasks. Effective learning from automatically collected mixed-quality data enhances data efficiency and model performance. Extensive experiments demonstrate that MoRE outperforms all baselines across six different skills and exhibits superior generalization capabilities in out-of-distribution scenarios. We further validate our method in real-world scenarios, confirming the practicality of our approach and laying a solid foundation for future research on multi-task learning in quadruped robots.

View on arXiv
@article{zhao2025_2503.08007,
  title={ MoRE: Unlocking Scalability in Reinforcement Learning for Quadruped Vision-Language-Action Models },
  author={ Han Zhao and Wenxuan Song and Donglin Wang and Xinyang Tong and Pengxiang Ding and Xuelian Cheng and Zongyuan Ge },
  journal={arXiv preprint arXiv:2503.08007},
  year={ 2025 }
}
Comments on this paper