We introduce MD-ProjTex, a method for fast and consistent text-guided texture generation for 3D shapes using pretrained text-to-image diffusion models. At the core of our approach is a multi-view consistency mechanism in UV space, which ensures coherent textures across different viewpoints. Specifically, MD-ProjTex fuses noise predictions from multiple views at each diffusion step and jointly updates the per-view denoising directions to maintain 3D consistency. In contrast to existing state-of-the-art methods that rely on optimization or sequential view synthesis, MD-ProjTex is computationally more efficient and achieves better quantitative and qualitative results.
View on arXiv@article{yildirim2025_2504.02762, title={ MD-ProjTex: Texturing 3D Shapes with Multi-Diffusion Projection }, author={ Ahmet Burak Yildirim and Mustafa Utku Aydogdu and Duygu Ceylan and Aysegul Dundar }, journal={arXiv preprint arXiv:2504.02762}, year={ 2025 } }