4
0

FlowR: Flowing from Sparse to Dense 3D Reconstructions

Abstract

3D Gaussian splatting enables high-quality novel view synthesis (NVS) at real-time frame rates. However, its quality drops sharply as we depart from the training views. Thus, dense captures are needed to match the high-quality expectations of some applications, e.g. Virtual Reality (VR). However, such dense captures are very laborious and expensive to obtain. Existing works have explored using 2D generative models to alleviate this requirement by distillation or generating additional training views. These methods are often conditioned only on a handful of reference input views and thus do not fully exploit the available 3D information, leading to inconsistent generation results and reconstruction artifacts. To tackle this problem, we propose a multi-view, flow matching model that learns a flow to connect novel view renderings from possibly sparse reconstructions to renderings that we expect from dense reconstructions. This enables augmenting scene captures with novel, generated views to improve reconstruction quality. Our model is trained on a novel dataset of 3.6M image pairs and can process up to 45 views at 540x960 resolution (91K tokens) on one H100 GPU in a single forward pass. Our pipeline consistently improves NVS in sparse- and dense-view scenarios, leading to higher-quality reconstructions than prior works across multiple, widely-used NVS benchmarks.

View on arXiv
@article{fischer2025_2504.01647,
  title={ FlowR: Flowing from Sparse to Dense 3D Reconstructions },
  author={ Tobias Fischer and Samuel Rota Bulò and Yung-Hsu Yang and Nikhil Varma Keetha and Lorenzo Porzi and Norman Müller and Katja Schwarz and Jonathon Luiten and Marc Pollefeys and Peter Kontschieder },
  journal={arXiv preprint arXiv:2504.01647},
  year={ 2025 }
}
Comments on this paper