14
0

UMoE: Unifying Attention and FFN with Shared Experts

Abstract

Sparse Mixture of Experts (MoE) architectures have emerged as a promising approach for scaling Transformer models. While initial works primarily incorporated MoE into feed-forward network (FFN) layers, recent studies have explored extending the MoE paradigm to attention layers to enhance model performance. However, existing attention-based MoE layers require specialized implementations and demonstrate suboptimal performance compared to their FFN-based counterparts. In this paper, we aim to unify the MoE designs in attention and FFN layers by introducing a novel reformulation of the attention mechanism, revealing an underlying FFN-like structure within attention modules. Our proposed architecture, UMoE, achieves superior performance through attention-based MoE layers while enabling efficient parameter sharing between FFN and attention components.

View on arXiv
@article{yang2025_2505.07260,
  title={ UMoE: Unifying Attention and FFN with Shared Experts },
  author={ Yuanhang Yang and Chaozheng Wang and Jing Li },
  journal={arXiv preprint arXiv:2505.07260},
  year={ 2025 }
}
Comments on this paper