ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2504.13206
41
0

DuoLoRA : Cycle-consistent and Rank-disentangled Content-Style Personalization

15 April 2025
Aniket Roy
Shubhankar Borse
Shreya Kadambi
Debasmit Das
Shweta Mahajan
Risheek Garrepalli
Hyojin Park
Ankita Nayak
Rama Chellappa
Munawar Hayat
Fatih Porikli
    MoMe
ArXivPDFHTML
Abstract

We tackle the challenge of jointly personalizing content and style from a few examples. A promising approach is to train separate Low-Rank Adapters (LoRA) and merge them effectively, preserving both content and style. Existing methods, such as ZipLoRA, treat content and style as independent entities, merging them by learning masks in LoRA's output dimensions. However, content and style are intertwined, not independent. To address this, we propose DuoLoRA, a content-style personalization framework featuring three key components: (i) rank-dimension mask learning, (ii) effective merging via layer priors, and (iii) Constyle loss, which leverages cycle-consistency in the merging process. First, we introduce ZipRank, which performs content-style merging within the rank dimension, offering adaptive rank flexibility and significantly reducing the number of learnable parameters. Additionally, we incorporate SDXL layer priors to apply implicit rank constraints informed by each layer's content-style bias and adaptive merger initialization, enhancing the integration of content and style. To further refine the merging process, we introduce Constyle loss, which leverages the cycle-consistency between content and style. Our experimental results demonstrate that DuoLoRA outperforms state-of-the-art content-style merging methods across multiple benchmarks.

View on arXiv
@article{roy2025_2504.13206,
  title={ DuoLoRA : Cycle-consistent and Rank-disentangled Content-Style Personalization },
  author={ Aniket Roy and Shubhankar Borse and Shreya Kadambi and Debasmit Das and Shweta Mahajan and Risheek Garrepalli and Hyojin Park and Ankita Nayak and Rama Chellappa and Munawar Hayat and Fatih Porikli },
  journal={arXiv preprint arXiv:2504.13206},
  year={ 2025 }
}
Comments on this paper