ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2212.08066
  4. Cited By
Mod-Squad: Designing Mixture of Experts As Modular Multi-Task Learners

Mod-Squad: Designing Mixture of Experts As Modular Multi-Task Learners

15 December 2022
Zitian Chen
Songlin Yang
Mingyu Ding
Zhenfang Chen
Hengshuang Zhao
E. Learned-Miller
Chuang Gan
    MoE
ArXiv (abs)PDFHTMLGithub

Papers citing "Mod-Squad: Designing Mixture of Experts As Modular Multi-Task Learners"

7 / 7 papers shown
Point-MoE: Large-Scale Multi-Dataset Training with Mixture-of-Experts for 3D Semantic Segmentation
Point-MoE: Large-Scale Multi-Dataset Training with Mixture-of-Experts for 3D Semantic Segmentation
Xuweiyi Chen
Wentao Zhou
Aruni RoyChowdhury
Zezhou Cheng
3DPC
433
2
0
29 May 2025
Towards Modular LLMs by Building and Reusing a Library of LoRAs
Towards Modular LLMs by Building and Reusing a Library of LoRAsInternational Conference on Machine Learning (ICML), 2024
O. Ostapenko
Zhan Su
Edoardo Ponti
Laurent Charlin
Nicolas Le Roux
Matheus Pereira
Lucas Caccia
Alessandro Sordoni
MoMe
283
68
0
18 May 2024
Bridging Remote Sensors with Multisensor Geospatial Foundation Models
Bridging Remote Sensors with Multisensor Geospatial Foundation Models
Boran Han
Shuai Zhang
Xingjian Shi
Markus Reichstein
327
56
0
01 Apr 2024
DiffusionMTL: Learning Multi-Task Denoising Diffusion Model from
  Partially Annotated Data
DiffusionMTL: Learning Multi-Task Denoising Diffusion Model from Partially Annotated DataComputer Vision and Pattern Recognition (CVPR), 2024
Hanrong Ye
Dan Xu
DiffM
314
20
0
22 Mar 2024
Task-Based MoE for Multitask Multilingual Machine Translation
Task-Based MoE for Multitask Multilingual Machine Translation
Hai Pham
Young Jin Kim
Subhabrata Mukherjee
David P. Woodruff
Barnabás Póczós
Hany Awadalla
MoE
303
8
0
30 Aug 2023
Experts Weights Averaging: A New General Training Scheme for Vision
  Transformers
Experts Weights Averaging: A New General Training Scheme for Vision Transformers
Yongqian Huang
Peng Ye
Xiaoshui Huang
Sheng Li
Tao Chen
Tong He
Wanli Ouyang
MoMe
256
10
0
11 Aug 2023
SILO Language Models: Isolating Legal Risk In a Nonparametric Datastore
SILO Language Models: Isolating Legal Risk In a Nonparametric DatastoreInternational Conference on Learning Representations (ICLR), 2023
Sewon Min
Suchin Gururangan
Eric Wallace
Hannaneh Hajishirzi
Noah A. Smith
Luke Zettlemoyer
AILaw
339
91
0
08 Aug 2023
1
Page 1 of 1