ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2505.13982
25
0

Adaptive Visuo-Tactile Fusion with Predictive Force Attention for Dexterous Manipulation

20 May 2025
Jinzhou Li
Tianhao Wu
Jiyao Zhang
Zeyuan Chen
Haotian Jin
Mingdong Wu
Yujun Shen
Yaodong Yang
Hao Dong
ArXivPDFHTML
Abstract

Effectively utilizing multi-sensory data is important for robots to generalize across diverse tasks. However, the heterogeneous nature of these modalities makes fusion challenging. Existing methods propose strategies to obtain comprehensively fused features but often ignore the fact that each modality requires different levels of attention at different manipulation stages. To address this, we propose a force-guided attention fusion module that adaptively adjusts the weights of visual and tactile features without human labeling. We also introduce a self-supervised future force prediction auxiliary task to reinforce the tactile modality, improve data imbalance, and encourage proper adjustment. Our method achieves an average success rate of 93% across three fine-grained, contactrich tasks in real-world experiments. Further analysis shows that our policy appropriately adjusts attention to each modality at different manipulation stages. The videos can be viewed atthis https URL.

View on arXiv
@article{li2025_2505.13982,
  title={ Adaptive Visuo-Tactile Fusion with Predictive Force Attention for Dexterous Manipulation },
  author={ Jinzhou Li and Tianhao Wu and Jiyao Zhang and Zeyuan Chen and Haotian Jin and Mingdong Wu and Yujun Shen and Yaodong Yang and Hao Dong },
  journal={arXiv preprint arXiv:2505.13982},
  year={ 2025 }
}
Comments on this paper