Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2404.09977
Cited By
MaxFusion: Plug&Play Multi-Modal Generation in Text-to-Image Diffusion Models
15 April 2024
Nithin Gopalakrishnan Nair
Jeya Maria Jose Valanarasu
Vishal M. Patel
MoMe
Re-assign community
ArXiv
PDF
HTML
Papers citing
"MaxFusion: Plug&Play Multi-Modal Generation in Text-to-Image Diffusion Models"
8 / 8 papers shown
Title
DMM: Building a Versatile Image Generation Model via Distillation-Based Model Merging
Tianhui Song
Weixin Feng
Shuai Wang
X. Li
Tiezheng Ge
Bo Zheng
Limin Wang
MoMe
57
0
0
16 Apr 2025
PartStickers: Generating Parts of Objects for Rapid Prototyping
Mo Zhou
Josh Myers-Dean
Danna Gurari
21
0
0
07 Apr 2025
Unconditional Priors Matter! Improving Conditional Generation of Fine-Tuned Diffusion Models
Prin Phunyaphibarn
Phillip Y. Lee
Jaihoon Kim
Minhyuk Sung
DiffM
78
0
0
26 Mar 2025
Text-DiFuse: An Interactive Multi-Modal Image Fusion Framework based on Text-modulated Diffusion Model
Hao Zhang
Lei Cao
Jiayi Ma
DiffM
33
4
0
31 Oct 2024
MagicBrush: A Manually Annotated Dataset for Instruction-Guided Image Editing
Kai Zhang
Lingbo Mo
Wenhu Chen
Huan Sun
Yu-Chuan Su
EGVM
105
235
0
16 Jun 2023
Git Re-Basin: Merging Models modulo Permutation Symmetries
Samuel K. Ainsworth
J. Hayase
S. Srinivasa
MoMe
239
313
0
11 Sep 2022
Pretraining is All You Need for Image-to-Image Translation
Tengfei Wang
Ting Zhang
Bo Zhang
Hao Ouyang
Dong Chen
Qifeng Chen
Fang Wen
DiffM
176
177
0
25 May 2022
BLIP: Bootstrapping Language-Image Pre-training for Unified Vision-Language Understanding and Generation
Junnan Li
Dongxu Li
Caiming Xiong
S. Hoi
MLLM
BDL
VLM
CLIP
382
4,010
0
28 Jan 2022
1