11
0

Learning Graph Representation of Agent Diffuser

Abstract

Diffusion-based generative models have significantly advanced text-to-image synthesis, demonstrating impressive text comprehension and zero-shot generalization. These models refine images from random noise based on textual prompts, with initial reliance on text input shifting towards enhanced visual fidelity over time. This transition suggests that static model parameters might not optimally address the distinct phases of generation. We introduce LGR-AD (Learning Graph Representation of Agent Diffusers), a novel multi-agent system designed to improve adaptability in dynamic computer vision tasks. LGR-AD models the generation process as a distributed system of interacting agents, each representing an expert sub-model. These agents dynamically adapt to varying conditions and collaborate through a graph neural network that encodes their relationships and performance metrics. Our approach employs a coordination mechanism based on top-kk maximum spanning trees, optimizing the generation process. Each agent's decision-making is guided by a meta-model that minimizes a novel loss function, balancing accuracy and diversity. Theoretical analysis and extensive empirical evaluations show that LGR-AD outperforms traditional diffusion models across various benchmarks, highlighting its potential for scalable and flexible solutions in complex image generation tasks. Code is available at:this https URL

View on arXiv
@article{djenouri2025_2505.06761,
  title={ Learning Graph Representation of Agent Diffuser },
  author={ Youcef Djenouri and Nassim Belmecheri and Tomasz Michalak and Jan Dubiński and Ahmed Nabil Belbachir and Anis Yazidi },
  journal={arXiv preprint arXiv:2505.06761},
  year={ 2025 }
}
Comments on this paper