6
v1v2 (latest)

NCCL EP: Towards a Unified Expert Parallel Communication API for NCCL

Amos Goldman
Nimrod Boker
Maayan Sheraizin
Nimrod Admoni
Artem Polyakov
Subhadeep Bhattacharya
Fan Yu
Kai Sun
Georgios Theodorakis
Hsin-Chun Yin
Peter-Jan Gootzen
Aamir Shafi
Assaf Ravid
Salvatore Di Girolamo
Manjunath Gorentla Venkata
Gil Bloch
Main:12 Pages
8 Figures
Bibliography:1 Pages
Abstract

Mixture-of-Experts (MoE) architectures have become essential for scaling large language models, driving the development of specialized device-initiated communication libraries such as DeepEP, Hybrid-EP, and others. These libraries demonstrate the performance benefits of GPU-initiated RDMA for MoE dispatch and combine operations.This paper presents NCCL EP (Expert Parallelism), a ground-up MoE communication library built entirely on NCCL's Device API. NCCL EP provides unified ncclEpDispatch and ncclEpCombine primitives with both C and Python interfaces, supporting Low-Latency (LL) mode for inference decoding and High-Throughput (HT) mode for training and inference prefill. LL targets small batch sizes (1-128 tokens) using direct all-to-all RDMA+NVLink mesh connectivity with double-buffered communication for overlapping dispatch and combine phases. HT targets large batches (4096+ tokens) using hierarchical communication that aggregates tokens within NVLink domains before inter-node RDMA transmission. Both modes leverage Device API for both intra- and inter-node communications, taking advantage of its topology awareness and optimized GPU-initiated implementation.We evaluate NCCL EP on an H100-based cluster across multi-node configurations, demonstrating competitive LL kernel performance and presenting end-to-end results with vLLM integration. By building MoE communication natively within NCCL, NCCL EP provides a supported path for expert parallelism on current and emerging NVIDIA platforms.

View on arXiv
Comments on this paper