317
v1v2 (latest)

SkillMimic: Learning Basketball Interaction Skills from Demonstrations

Computer Vision and Pattern Recognition (CVPR), 2024
Yinhuai Wang
Qifeng Chen
Jian Zhang
Lei Zhang
Ping Tan
Main:8 Pages
15 Figures
Bibliography:3 Pages
9 Tables
Appendix:7 Pages
Abstract

Traditional reinforcement learning methods for human-object interaction (HOI) rely on labor-intensive, manually designed skill rewards that do not generalize well across different interactions. We introduce SkillMimic, a unified data-driven framework that fundamentally changes how agents learn interaction skills by eliminating the need for skill-specific rewards. Our key insight is that a unified HOI imitation reward can effectively capture the essence of diverse interaction patterns from HOI datasets. This enables SkillMimic to learn a single policy that not only masters multiple interaction skills but also facilitates skill transitions, with both diversity and generalization improving as the HOI dataset grows. For evaluation, we collect and introduce two basketball datasets containing approximately 35 minutes of diverse basketball skills. Extensive experiments show that SkillMimic successfully masters a wide range of basketball skills including stylistic variations in dribbling, layup, and shooting. Moreover, these learned skills can be effectively composed by a high-level controller to accomplish complex and long-horizon tasks such as consecutive scoring, opening new possibilities for scalable and generalizable interaction skill learning. Project page:this https URL

View on arXiv
Comments on this paper