73
v1v2 (latest)

Diffusion Timbre Transfer Via Mutual Information Guided Inpainting

Ching Ho Lee
Javier Nistal
Stefan Lattner
Marco Pasini
George Fazekas
Main:3 Pages
2 Figures
Bibliography:2 Pages
3 Tables
Abstract

We study timbre transfer as an inference-time editing problem for music audio. Starting from a strong pre-trained latent diffusion model, we introduce a lightweight procedure that requires no additional training: (i) a dimension-wise noise injection that targets latent channels most informative of instrument identity, and (ii) an early-step clamping mechanism that re-imposes the input's melodic and rhythmic structure during reverse diffusion. The method operates directly on audio latents and is compatible with text/audio conditioning (e.g., CLAP). We discuss design choices,analyze trade-offs between timbral change and structural preservation, and show that simple inference-time controls can meaningfully steer pre-trained models for style-transfer use cases.

View on arXiv
Comments on this paper