ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2406.08920
25
5

AV-GS: Learning Material and Geometry Aware Priors for Novel View Acoustic Synthesis

13 June 2024
Swapnil Bhosale
Haosen Yang
Diptesh Kanojia
Jiankang Deng
Xiatian Zhu
ArXivPDFHTML
Abstract

Novel view acoustic synthesis (NVAS) aims to render binaural audio at any target viewpoint, given a mono audio emitted by a sound source at a 3D scene. Existing methods have proposed NeRF-based implicit models to exploit visual cues as a condition for synthesizing binaural audio. However, in addition to low efficiency originating from heavy NeRF rendering, these methods all have a limited ability of characterizing the entire scene environment such as room geometry, material properties, and the spatial relation between the listener and sound source. To address these issues, we propose a novel Audio-Visual Gaussian Splatting (AV-GS) model. To obtain a material-aware and geometry-aware condition for audio synthesis, we learn an explicit point-based scene representation with an audio-guidance parameter on locally initialized Gaussian points, taking into account the space relation from the listener and sound source. To make the visual scene model audio adaptive, we propose a point densification and pruning strategy to optimally distribute the Gaussian points, with the per-point contribution in sound propagation (e.g., more points needed for texture-less wall surfaces as they affect sound path diversion). Extensive experiments validate the superiority of our AV-GS over existing alternatives on the real-world RWAS and simulation-based SoundSpaces datasets.

View on arXiv
@article{bhosale2025_2406.08920,
  title={ AV-GS: Learning Material and Geometry Aware Priors for Novel View Acoustic Synthesis },
  author={ Swapnil Bhosale and Haosen Yang and Diptesh Kanojia and Jiankang Deng and Xiatian Zhu },
  journal={arXiv preprint arXiv:2406.08920},
  year={ 2025 }
}
Comments on this paper