ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2503.04215
60
0

Energy-Guided Optimization for Personalized Image Editing with Pretrained Text-to-Image Diffusion Models

6 March 2025
Rui Jiang
Xinghe Fu
Guangcong Zheng
Teng Li
Taiping Yao
Xi Li
    DiffM
ArXivPDFHTML
Abstract

The rapid advancement of pretrained text-driven diffusion models has significantly enriched applications in image generation and editing. However, as the demand for personalized content editing increases, new challenges emerge especially when dealing with arbitrary objects and complex scenes. Existing methods usually mistakes mask as the object shape prior, which struggle to achieve a seamless integration result. The mostly used inversion noise initialization also hinders the identity consistency towards the target object. To address these challenges, we propose a novel training-free framework that formulates personalized content editing as the optimization of edited images in the latent space, using diffusion models as the energy function guidance conditioned by reference text-image pairs. A coarse-to-fine strategy is proposed that employs text energy guidance at the early stage to achieve a natural transition toward the target class and uses point-to-point feature-level image energy guidance to perform fine-grained appearance alignment with the target object. Additionally, we introduce the latent space content composition to enhance overall identity consistency with the target. Extensive experiments demonstrate that our method excels in object replacement even with a large domain gap, highlighting its potential for high-quality, personalized image editing.

View on arXiv
@article{jiang2025_2503.04215,
  title={ Energy-Guided Optimization for Personalized Image Editing with Pretrained Text-to-Image Diffusion Models },
  author={ Rui Jiang and Xinghe Fu and Guangcong Zheng and Teng Li and Taiping Yao and Xi Li },
  journal={arXiv preprint arXiv:2503.04215},
  year={ 2025 }
}
Comments on this paper