ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2506.09997
56
0

DGS-LRM: Real-Time Deformable 3D Gaussian Reconstruction From Monocular Videos

11 June 2025
C. Lin
Zhaoyang Lv
Songyin Wu
Zhen Xu
Thu Nguyen-Phuoc
Hung-Yu Tseng
Julian Straub
Numair Khan
Lei Xiao
Ming-Hsuan Yang
Yuheng Ren
Richard Newcombe
Zhao Dong
Zhengqin Li
    3DGS
ArXiv (abs)PDFHTML
Main:8 Pages
7 Figures
Bibliography:7 Pages
3 Tables
Abstract

We introduce the Deformable Gaussian Splats Large Reconstruction Model (DGS-LRM), the first feed-forward method predicting deformable 3D Gaussian splats from a monocular posed video of any dynamic scene. Feed-forward scene reconstruction has gained significant attention for its ability to rapidly create digital replicas of real-world environments. However, most existing models are limited to static scenes and fail to reconstruct the motion of moving objects. Developing a feed-forward model for dynamic scene reconstruction poses significant challenges, including the scarcity of training data and the need for appropriate 3D representations and training paradigms. To address these challenges, we introduce several key technical contributions: an enhanced large-scale synthetic dataset with ground-truth multi-view videos and dense 3D scene flow supervision; a per-pixel deformable 3D Gaussian representation that is easy to learn, supports high-quality dynamic view synthesis, and enables long-range 3D tracking; and a large transformer network that achieves real-time, generalizable dynamic scene reconstruction. Extensive qualitative and quantitative experiments demonstrate that DGS-LRM achieves dynamic scene reconstruction quality comparable to optimization-based methods, while significantly outperforming the state-of-the-art predictive dynamic reconstruction method on real-world examples. Its predicted physically grounded 3D deformation is accurate and can readily adapt for long-range 3D tracking tasks, achieving performance on par with state-of-the-art monocular video 3D tracking methods.

View on arXiv
@article{lin2025_2506.09997,
  title={ DGS-LRM: Real-Time Deformable 3D Gaussian Reconstruction From Monocular Videos },
  author={ Chieh Hubert Lin and Zhaoyang Lv and Songyin Wu and Zhen Xu and Thu Nguyen-Phuoc and Hung-Yu Tseng and Julian Straub and Numair Khan and Lei Xiao and Ming-Hsuan Yang and Yuheng Ren and Richard Newcombe and Zhao Dong and Zhengqin Li },
  journal={arXiv preprint arXiv:2506.09997},
  year={ 2025 }
}
Comments on this paper