ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2512.00877
0
0

Feed-Forward 3D Gaussian Splatting Compression with Long-Context Modeling

30 November 2025
Zhening Liu
Rui Song
Yushi Huang
Yingdong Hu
Xinjie Zhang
Jiawei Shao
Zehong Lin
Jun Zhang
    3DGS
ArXiv (abs)PDFHTML
Main:8 Pages
18 Figures
Bibliography:3 Pages
6 Tables
Appendix:6 Pages
Abstract

3D Gaussian Splatting (3DGS) has emerged as a revolutionary 3D representation. However, its substantial data size poses a major barrier to widespread adoption. While feed-forward 3DGS compression offers a practical alternative to costly per-scene per-train compressors, existing methods struggle to model long-range spatial dependencies, due to the limited receptive field of transform coding networks and the inadequate context capacity in entropy models. In this work, we propose a novel feed-forward 3DGS compression framework that effectively models long-range correlations to enable highly compact and generalizable 3D representations. Central to our approach is a large-scale context structure that comprises thousands of Gaussians based on Morton serialization. We then design a fine-grained space-channel auto-regressive entropy model to fully leverage this expansive context. Furthermore, we develop an attention-based transform coding model to extract informative latent priors by aggregating features from a wide range of neighboring Gaussians. Our method yields a 20×20\times20× compression ratio for 3DGS in a feed-forward inference and achieves state-of-the-art performance among generalizable codecs.

View on arXiv
Comments on this paper