D^2USt3R: Enhancing 3D Reconstruction with 4D Pointmaps for Dynamic Scenes

We address the task of 3D reconstruction in dynamic scenes, where object motions degrade the quality of previous 3D pointmap regression methods, such as DUSt3R, originally designed for static 3D scene reconstruction. Although these methods provide an elegant and powerful solution in static settings, they struggle in the presence of dynamic motions that disrupt alignment based solely on camera poses. To overcome this, we propose D^2USt3R that regresses 4D pointmaps that simultaneiously capture both static and dynamic 3D scene geometry in a feed-forward manner. By explicitly incorporating both spatial and temporal aspects, our approach successfully encapsulates spatio-temporal dense correspondence to the proposed 4D pointmaps, enhancing downstream tasks. Extensive experimental evaluations demonstrate that our proposed approach consistently achieves superior reconstruction performance across various datasets featuring complex motions.
View on arXiv@article{han2025_2504.06264, title={ D^2USt3R: Enhancing 3D Reconstruction with 4D Pointmaps for Dynamic Scenes }, author={ Jisang Han and Honggyu An and Jaewoo Jung and Takuya Narihira and Junyoung Seo and Kazumi Fukuda and Chaehyun Kim and Sunghwan Hong and Yuki Mitsufuji and Seungryong Kim }, journal={arXiv preprint arXiv:2504.06264}, year={ 2025 } }