Making Sense of Vision and Touch: Self-Supervised Learning of Multimodal
  Representations for Contact-Rich Tasks
v1v2 (latest)

Making Sense of Vision and Touch: Self-Supervised Learning of Multimodal Representations for Contact-Rich Tasks

    SSL

Papers citing "Making Sense of Vision and Touch: Self-Supervised Learning of Multimodal Representations for Contact-Rich Tasks"

50 / 219 papers shown
Title
Contrastive Touch-to-Touch Pretraining
Contrastive Touch-to-Touch PretrainingIEEE International Conference on Robotics and Automation (ICRA), 2024
158
4
0
15 Oct 2024
RDT-1B: a Diffusion Foundation Model for Bimanual Manipulation
RDT-1B: a Diffusion Foundation Model for Bimanual ManipulationInternational Conference on Learning Representations (ICLR), 2024
247
282
0
10 Oct 2024
Hearing Touch: Audio-Visual Pretraining for Contact-Rich Manipulation
Hearing Touch: Audio-Visual Pretraining for Contact-Rich ManipulationIEEE International Conference on Robotics and Automation (ICRA), 2024
155
17
0
14 May 2024
Cross-domain Multi-modal Few-shot Object Detection via Rich Text
Cross-domain Multi-modal Few-shot Object Detection via Rich TextIEEE Workshop/Winter Conference on Applications of Computer Vision (WACV), 2024
254
1
0
24 Mar 2024

We use cookies and other tracking technologies to improve your browsing experience on our website, to show you personalized content and targeted ads, to analyze our website traffic, and to understand where our visitors are coming from. See our policy.