ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2502.15542
54
0

Bridging Domain Gaps between Pretrained Multimodal Models and Recommendations

24 February 2025
Wenyu Zhang
Jie Luo
Xinming Zhang
Yuan Fang
ArXivPDFHTML
Abstract

With the explosive growth of multimodal content online, pre-trained visual-language models have shown great potential for multimodal recommendation. However, while these models achieve decent performance when applied in a frozen manner, surprisingly, due to significant domain gaps (e.g., feature distribution discrepancy and task objective misalignment) between pre-training and personalized recommendation, adopting a joint training approach instead leads to performance worse than baseline. Existing approaches either rely on simple feature extraction or require computationally expensive full model fine-tuning, struggling to balance effectiveness and efficiency. To tackle these challenges, we propose \textbf{P}arameter-efficient \textbf{T}uning for \textbf{M}ultimodal \textbf{Rec}ommendation (\textbf{PTMRec}), a novel framework that bridges the domain gap between pre-trained models and recommendation systems through a knowledge-guided dual-stage parameter-efficient training strategy. This framework not only eliminates the need for costly additional pre-training but also flexibly accommodates various parameter-efficient tuning methods.

View on arXiv
@article{zhang2025_2502.15542,
  title={ Bridging Domain Gaps between Pretrained Multimodal Models and Recommendations },
  author={ Wenyu Zhang and Jie Luo and Xinming Zhang and Yuan Fang },
  journal={arXiv preprint arXiv:2502.15542},
  year={ 2025 }
}
Comments on this paper