v1v2 (latest)
Flexible and Effective Mixing of Large Language Models into a Mixture of
Domain Experts
- MoE
Main:17 Pages
9 Figures
Bibliography:1 Pages
2 Tables
Appendix:1 Pages
Abstract
We present a toolkit for creating low-cost Mixture-of-Domain-Experts (MOE) from trained models. The toolkit can be used for creating a mixture from models or from adapters. We perform extensive tests and offer guidance on defining the architecture of the resulting MOE using the toolkit. A public repository is available.
View on arXivComments on this paper
