ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2503.07111
36
0

PoseLess: Depth-Free Vision-to-Joint Control via Direct Image Mapping with VLM

10 March 2025
Alan Dao
Dinh Bach Vu
Tuan Le Duc Anh
Bui Quang Huy
ArXivPDFHTML
Abstract

This paper introduces PoseLess, a novel framework for robot hand control that eliminates the need for explicit pose estimation by directly mapping 2D images to joint angles using projected representations. Our approach leverages synthetic training data generated through randomized joint configurations, enabling zero-shot generalization to real-world scenarios and cross-morphology transfer from robotic to human hands. By projecting visual inputs and employing a transformer-based decoder, PoseLess achieves robust, low-latency control while addressing challenges such as depth ambiguity and data scarcity. Experimental results demonstrate competitive performance in joint angle prediction accuracy without relying on any human-labelled dataset.

View on arXiv
@article{dao2025_2503.07111,
  title={ PoseLess: Depth-Free Vision-to-Joint Control via Direct Image Mapping with VLM },
  author={ Alan Dao and Dinh Bach Vu and Tuan Le Duc Anh and Bui Quang Huy },
  journal={arXiv preprint arXiv:2503.07111},
  year={ 2025 }
}
Comments on this paper