ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Communities
  3. ...

Neighbor communities

0 / 0 papers shown
Top Contributors
Name# Papers# Citations
Social Events
DateLocationEvent
  1. Home
  2. Communities
  3. MoE

Mixture of Experts

MoE
More data

Mixture of Experts (MoE) is a machine learning technique that uses multiple expert models to make predictions. Each expert specializes in different aspects of the data, and a gating network determines which expert to use for a given input. This approach can improve model performance and efficiency.

Neighbor communities

51015

Featured Papers

0 / 0 papers shown

All papers

50 / 1,625 papers shown
Towards Principled Design of Mixture-of-Experts Language Models under Memory and Inference Constraints
Towards Principled Design of Mixture-of-Experts Language Models under Memory and Inference Constraints
Seng Pei Liew
Kenta Shinzato
Yuyang Dong
MoE
64
0
0
13 Jan 2026
Ministral 3
Ministral 3
Alexander H. Liu
Kartik Khandelwal
Sandeep Subramanian
Victor Jouault
Abhinav Rastogi
...
Thibaut Lavril
Thiziri Nait Saada
Thomas Chabal
Thomas Foubert
Thomas Robert
MoE
8
0
0
13 Jan 2026
Deconstructing Pre-training: Knowledge Attribution Analysis in MoE and Dense Models
Deconstructing Pre-training: Knowledge Attribution Analysis in MoE and Dense Models
Bo Wang
Junzhuo Li
Hong Chen
Yuanlin Chu
Yuxuan Fan
Xuming Hu
MoE
0
0
0
13 Jan 2026
Conditional Memory via Scalable Lookup: A New Axis of Sparsity for Large Language Models
Conditional Memory via Scalable Lookup: A New Axis of Sparsity for Large Language Models
Xin Cheng
Wangding Zeng
Damai Dai
Qinyu Chen
Bingxuan Wang
...
Yukun Li
Han Zhang
Huishuai Zhang
Dongyan Zhao
Wenfeng Liang
MoE
8
0
0
12 Jan 2026
Towards Specialized Generalists: A Multi-Task MoE-LoRA Framework for Domain-Specific LLM Adaptation
Towards Specialized Generalists: A Multi-Task MoE-LoRA Framework for Domain-Specific LLM Adaptation
Yuxin Yang
Aoxiong Zeng
Xiangquan Yang
MoMeMoE
88
0
0
12 Jan 2026
High-Rank Structured Modulation for Parameter-Efficient Fine-Tuning
High-Rank Structured Modulation for Parameter-Efficient Fine-Tuning
Yongkang Liu
Xing Li
Mengjie Zhao
Shanru Zhang
Zijing Wang
Qian Li
Shi Feng
Feiliang Ren
Daling Wang
Hinrich Schütze
MoE
0
0
0
12 Jan 2026
SecMoE: Communication-Efficient Secure MoE Inference via Select-Then-Compute
SecMoE: Communication-Efficient Secure MoE Inference via Select-Then-Compute
Bowen Shen
Yuyue Chen
Peng Yang
Bin Zhang
Xi Zhang
Zoe L. Jiang
MoE
100
0
0
11 Jan 2026
MoE-DisCo:Low Economy Cost Training Mixture-of-Experts Models
MoE-DisCo:Low Economy Cost Training Mixture-of-Experts Models
Xin Ye
Daning Cheng
Boyang Zhang
Yunquan Zhang
MoE
0
0
0
11 Jan 2026
Monkey Jump : MoE-Style PEFT for Efficient Multi-Task Learning
Monkey Jump : MoE-Style PEFT for Efficient Multi-Task Learning
Nusrat Jahan Prottasha
Md Kowsher
Chun-Nam Yu
Chen Chen
Ozlem Garibay
MoE
0
0
0
09 Jan 2026
DR-LoRA: Dynamic Rank LoRA for Mixture-of-Experts Adaptation
DR-LoRA: Dynamic Rank LoRA for Mixture-of-Experts Adaptation
Guanzhi Deng
Bo Li
Ronghao Chen
Huacan Wang
Linqi Song
Lijie Wen
MoE
57
0
0
08 Jan 2026
MoE3D: A Mixture-of-Experts Module for 3D Reconstruction
MoE3D: A Mixture-of-Experts Module for 3D Reconstruction
Zichen Wang
Ang Cao
Liam J. Wang
Jeong Joon Park
MoE
73
0
0
08 Jan 2026
Variational Inference, Entropy, and Orthogonality: A Unified Theory of Mixture-of-Experts
Variational Inference, Entropy, and Orthogonality: A Unified Theory of Mixture-of-Experts
Ye Su
Yong Liu
MoE
4
0
0
07 Jan 2026
Spectral Manifold Regularization for Stable and Modular Routing in Deep MoE Architectures
Spectral Manifold Regularization for Stable and Modular Routing in Deep MoE Architectures
Ibrahim Delibasoglu
MoE
37
0
0
07 Jan 2026
A Scheduling Framework for Efficient MoE Inference on Edge GPU-NDP Systems
A Scheduling Framework for Efficient MoE Inference on Edge GPU-NDP Systems
Qi Wu
Chao Fang
Jiayuan Chen
Ye Lin
Yueqi Zhang
Yichuan Bai
Yuan Du
Li Du
MoE
4
0
0
07 Jan 2026
MiMo-V2-Flash Technical Report
MiMo-V2-Flash Technical Report
Xiaomi LLM-Core Team
Bangjun Xiao
Bingquan Xia
Bo Yang
Bofei Gao
...
Shicheng Li
Shuhao Gu
Shuhuai Ren
Sirui Deng
Tao Guo
MoEVLM
158
0
0
06 Jan 2026
The Illusion of Specialization: Unveiling the Domain-Invariant "Standing Committee" in Mixture-of-Experts Models
The Illusion of Specialization: Unveiling the Domain-Invariant "Standing Committee" in Mixture-of-Experts Models
Yan Wang
Yitao Xu
Nanhan Shen
Jinyan Su
Jimin Huang
Zining Zhu
MoE
12
0
0
06 Jan 2026
MoE Adapter for Large Audio Language Models: Sparsity, Disentanglement, and Gradient-Conflict-Free
MoE Adapter for Large Audio Language Models: Sparsity, Disentanglement, and Gradient-Conflict-Free
Yishu Lei
Shuwei He
Jing Hu
Dan Zhang
Xianlong Luo
...
Rui Liu
Jingzhou He
Yu Sun
Hua Wu
Haifeng Wang
AuLLMMoE
60
0
0
06 Jan 2026
Placement Semantics for Distributed Deep Learning: A Systematic Framework for Analyzing Parallelism Strategies
Placement Semantics for Distributed Deep Learning: A Systematic Framework for Analyzing Parallelism Strategies
Deep Pankajbhai Mehta
FedMLMoE
56
0
0
05 Jan 2026
K-EXAONE Technical Report
K-EXAONE Technical Report
Eunbi Choi
Kibong Choi
Seokhee Hong
Junwon Hwang
Hyojin Jeon
...
Sihyuk Yi
Chansik Yoon
Dongkeun Yoon
Sangyeon Yoon
Hyeongu Yun
MoEELM
228
0
0
05 Jan 2026
Routing by Analogy: kNN-Augmented Expert Assignment for Mixture-of-Experts
Routing by Analogy: kNN-Augmented Expert Assignment for Mixture-of-Experts
Boxuan Lyu
Soichiro Murakami
Hidetaka Kamigaito
Peinan Zhang
MoE
93
0
0
05 Jan 2026
Varying-Coefficient Mixture of Experts Model
Varying-Coefficient Mixture of Experts Model
Qicheng Zhao
Celia M.T. Greenwood
Qihuang Zhang
MoE
128
0
0
05 Jan 2026
Yuan3.0 Flash: An Open Multimodal Large Language Model for Enterprise Applications
Yuan3.0 Flash: An Open Multimodal Large Language Model for Enterprise Applications
YuanLab.ai
Shawn Wu
Sean Wang
Louie Li
Darcy Chen
...
James Gong
Danied Zhao
Penn Zheng
Owen Zhu
Tong Yu
MoELRM
196
0
0
05 Jan 2026
Making MoE-based LLM Inference Resilient with Tarragon
Making MoE-based LLM Inference Resilient with Tarragon
Songyu Zhang
Aaron Tam
Myungjin Lee
Shixiong Qi
K. K. Ramakrishnan
MoE
120
0
0
04 Jan 2026
Reliability Under Randomness: An Empirical Analysis of Sparse and Dense Language Models Across Decoding Temperatures
Reliability Under Randomness: An Empirical Analysis of Sparse and Dense Language Models Across Decoding Temperatures
Kabir Grover
MoE
0
0
0
02 Jan 2026
HFedMoE: Resource-aware Heterogeneous Federated Learning with Mixture-of-Experts
HFedMoE: Resource-aware Heterogeneous Federated Learning with Mixture-of-Experts
Zihan Fang
Zheng Lin
Senkang Hu
Yanan Ma
Yihang Tao
Yiqin Deng
Xianhao Chen
Yuguang Fang
MoE
56
0
0
02 Jan 2026
Traffic-MoE: A Sparse Foundation Model for Network Traffic Analysis
Traffic-MoE: A Sparse Foundation Model for Network Traffic Analysis
Jiajun Zhou
Changhui Sun
Meng Shen
Shanqing Yu
Qi Xuan
MoE
72
0
0
01 Jan 2026
RepetitionCurse: Measuring and Understanding Router Imbalance in Mixture-of-Experts LLMs under DoS Stress
RepetitionCurse: Measuring and Understanding Router Imbalance in Mixture-of-Experts LLMs under DoS Stress
Ruixuan Huang
Qingyue Wang
Hantao Huang
Yudong Gao
Dong Chen
Shuai Wang
Wei Wang
MoE
96
0
0
30 Dec 2025
Training Report of TeleChat3-MoE
Training Report of TeleChat3-MoE
Xinzhang Liu
Chao Wang
Zhihao Yang
Zhuo Jiang
Xuncheng Zhao
...
Teng Su
Xin Jiang
Shuangyong Song
Yongxiang Li
Xuelong Li
MoE
4
0
0
30 Dec 2025
Coupling Experts and Routers in Mixture-of-Experts via an Auxiliary Loss
Coupling Experts and Routers in Mixture-of-Experts via an Auxiliary Loss
Ang Lv
Jin Ma
Yiyuan Ma
Siyuan Qiao
MoE
65
0
0
29 Dec 2025
Text-Routed Sparse Mixture-of-Experts Model with Explanation and Temporal Alignment for Multi-Modal Sentiment Analysis
Text-Routed Sparse Mixture-of-Experts Model with Explanation and Temporal Alignment for Multi-Modal Sentiment Analysis
Dongning Rao
Yunbiao Zeng
Zhihua Jiang
Jujian Lv
MoE
8
0
0
28 Dec 2025
FLEX-MoE: Federated Mixture-of-Experts with Load-balanced Expert Assignment
FLEX-MoE: Federated Mixture-of-Experts with Load-balanced Expert Assignment
Boyang Zhang
Xiaobing Chen
Songyang Zhang
Shuai Zhang
Xiangwei Zhou
Mingxuan Sun
MoE
8
0
0
28 Dec 2025
Accelerate Speculative Decoding with Sparse Computation in Verification
Accelerate Speculative Decoding with Sparse Computation in Verification
Jikai Wang
Jianchao Tan
Yuxuan Hu
Jiayu Qin
Yerui Sun
Yuchen Xie
Xunliang Cai
Juntao Li
Min Zhang
MoE
24
0
0
26 Dec 2025
InstructMoLE: Instruction-Guided Mixture of Low-rank Experts for Multi-Conditional Image Generation
InstructMoLE: Instruction-Guided Mixture of Low-rank Experts for Multi-Conditional Image Generation
Jinqi Xiao
Qing Yan
Liming Jiang
Zichuan Liu
Hao Kang
...
Tiancheng Zhi
Jing Liu
Cheng Yang
Xin Lu
Bo Yuan
DiffMMoE
62
0
0
25 Dec 2025
Hybrid Quantum-Classical Mixture of Experts: Unlocking Topological Advantage via Interference-Based Routing
Hybrid Quantum-Classical Mixture of Experts: Unlocking Topological Advantage via Interference-Based Routing
Reda Heddad
Lamiae Bouanane
MoE
12
0
0
25 Dec 2025
MoRAgent: Parameter Efficient Agent Tuning with Mixture-of-Roles
MoRAgent: Parameter Efficient Agent Tuning with Mixture-of-RolesInternational Conference on Machine Learning (ICML), 2025
Jing Han
Binwei Yan
Tianyu Guo
Zheyuan Bai
Mengyu Zheng
Hanting Chen
Ying Nie
MoEOffRL
8
0
0
25 Dec 2025
Efficient MoE Inference with Fine-Grained Scheduling of Disaggregated Expert Parallelism
Efficient MoE Inference with Fine-Grained Scheduling of Disaggregated Expert Parallelism
Xinglin Pan
Shaohuai Shi
Wenxiang Lin
Yuxin Wang
Zhenheng Tang
Wei Wang
Xiaowen Chu
MoE
92
0
0
25 Dec 2025
RevFFN: Memory-Efficient Full-Parameter Fine-Tuning of Mixture-of-Experts LLMs with Reversible Blocks
RevFFN: Memory-Efficient Full-Parameter Fine-Tuning of Mixture-of-Experts LLMs with Reversible Blocks
Ningyuan Liu
Jing Yang
Kaitong Cai
Keze Wang
MoE
8
0
0
24 Dec 2025
Mixture of Experts in Large Language Models
Mixture of Experts in Large Language Models
Danyang Zhang
Junhao Song
Ziqian Bi
Xinyuan Song
Yingfang Yuan
Tianyang Wang
Joe Yeong
Junfeng Hao
MoE
139
0
0
24 Dec 2025
GateBreaker: Gate-Guided Attacks on Mixture-of-Expert LLMs
GateBreaker: Gate-Guided Attacks on Mixture-of-Expert LLMs
Lichao Wu
Sasha Behrouzi
Mohamadreza Rostami
Stjepan Picek
Ahmad-Reza Sadeghi
MoEAAML
272
0
0
24 Dec 2025
Insights into DeepSeek-V3: Scaling Challenges and Reflections on Hardware for AI Architectures
Insights into DeepSeek-V3: Scaling Challenges and Reflections on Hardware for AI ArchitecturesInternational Symposium on Computer Architecture (ISCA), 2025
Chenggang Zhao
Chengqi Deng
Chong Ruan
Damai Dai
Huazuo Gao
...
Wenfeng Liang
Ying He
Yun Wang
Yuxuan Liu
Y. X. Wei
MoE
243
33
0
24 Dec 2025
MoE-DiffuSeq: Enhancing Long-Document Diffusion Models with Sparse Attention and Mixture of Experts
MoE-DiffuSeq: Enhancing Long-Document Diffusion Models with Sparse Attention and Mixture of Experts
Alexandros Christoforos
Chadbourne Davis
DiffMMoE
0
0
0
23 Dec 2025
Mixture-of-Experts with Gradient Conflict-Driven Subspace Topology Pruning for Emergent Modularity
Mixture-of-Experts with Gradient Conflict-Driven Subspace Topology Pruning for Emergent Modularity
Yuxing Gan
Ziyu Lei
MoE
224
0
0
23 Dec 2025
Nemotron 3 Nano: Open, Efficient Mixture-of-Experts Hybrid Mamba-Transformer Model for Agentic Reasoning
Nemotron 3 Nano: Open, Efficient Mixture-of-Experts Hybrid Mamba-Transformer Model for Agentic Reasoning
NVIDIA
Aaron Blakeman
Aaron Grattafiori
Aarti Basant
Abhibha Gupta
...
Ferenc Galko
Frankie Siino
Gal Hubara Agam
Ganesh Ajjanagadde
Gantavya Bhatt
MoELRM
164
0
0
23 Dec 2025
Mirage Persistent Kernel: A Compiler and Runtime for Mega-Kernelizing Tensor Programs
Mirage Persistent Kernel: A Compiler and Runtime for Mega-Kernelizing Tensor Programs
Xinhao Cheng
Zhihao Zhang
Yu Zhou
Jianan Ji
Jinchen Jiang
...
Songting Wang
Wenqin Yang
Xupeng Miao
Tianqi Chen
Zhihao Jia
MoEVLM
12
0
0
22 Dec 2025
Small Language Models as Compiler Experts: Auto-Parallelization for Heterogeneous Systems
Small Language Models as Compiler Experts: Auto-Parallelization for Heterogeneous Systems
Prathamesh Devadiga
MoELRM
80
0
0
22 Dec 2025
Remoe: Towards Efficient and Low-Cost MoE Inference in Serverless Computing
Remoe: Towards Efficient and Low-Cost MoE Inference in Serverless Computing
Wentao Liu
Yuhao Hu
Ruiting Zhou
Baochun Li
Ne Wang
MoE
196
0
0
21 Dec 2025
Secret mixtures of experts inside your LLM
Secret mixtures of experts inside your LLM
Enric Boix-Adsera
MoE
20
0
0
20 Dec 2025
Sigma-MoE-Tiny Technical Report
Sigma-MoE-Tiny Technical Report
Qingguo Hu
Zhenghao Lin
Ziyue Yang
Yucheng Ding
Xiao Liu
...
Rui Gao
Lei Qu
Jinsong Su
Peng Cheng
Yeyun Gong
MoE
276
0
0
18 Dec 2025
LoPA: Scaling dLLM Inference via Lookahead Parallel Decoding
LoPA: Scaling dLLM Inference via Lookahead Parallel Decoding
Chenkai Xu
Yijie Jin
Jiajun Li
Yi Tu
Guoping Long
...
Mingcong Song
Hongjie Si
Tianqi Hou
Junchi Yan
Zhijie Deng
MoEAI4CE
176
0
0
18 Dec 2025
Bandwidth-Efficient Adaptive Mixture-of-Experts via Low-Rank Compensation
Bandwidth-Efficient Adaptive Mixture-of-Experts via Low-Rank Compensation
Zhenyu Liu
Yunzhen Liu
Zehao Fan
Garrett Gagnon
Yayue Hou
Nan Wu
Yangwook Kang
Liu Liu
MoE
188
0
0
18 Dec 2025
Loading #Papers per Month with "MoE"
Past speakers
Name (-)
Top Contributors
Name (-)
Top Organizations at ResearchTrend.AI
Name (-)
Social Events
DateLocationEvent
No social events available