Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2203.03312
Cited By
SkillNet-NLU: A Sparsely Activated Model for General-Purpose Natural Language Understanding
7 March 2022
Fan Zhang
Duyu Tang
Yong Dai
Cong Zhou
Shuangzhi Wu
Shuming Shi
CLL
MoE
Re-assign community
ArXiv
PDF
HTML
Papers citing
"SkillNet-NLU: A Sparsely Activated Model for General-Purpose Natural Language Understanding"
10 / 10 papers shown
Title
ScaLearn: Simple and Highly Parameter-Efficient Task Transfer by Learning to Scale
Markus Frohmann
Carolin Holtermann
Shahed Masoudian
Anne Lauscher
Navid Rekabsaz
29
2
0
02 Oct 2023
LoraHub: Efficient Cross-Task Generalization via Dynamic LoRA Composition
Chengsong Huang
Qian Liu
Bill Yuchen Lin
Tianyu Pang
Chao Du
Min-Bin Lin
MoMe
36
182
0
25 Jul 2023
SkillNet-X: A Multilingual Multitask Model with Sparsely Activated Skills
Zhangyin Feng
Yong Dai
Fan Zhang
Duyu Tang
Xiaocheng Feng
Shuangzhi Wu
Bing Qin
Yunbo Cao
Shuming Shi
MoE
29
0
0
28 Jun 2023
Modular Deep Learning
Jonas Pfeiffer
Sebastian Ruder
Ivan Vulić
E. Ponti
MoMe
OOD
21
73
0
22 Feb 2023
One Model for All Domains: Collaborative Domain-Prefix Tuning for Cross-Domain NER
Xiang Chen
Lei Li
Q. Fei
Ningyu Zhang
Chuanqi Tan
Yong-jia Jiang
Fei Huang
Huajun Chen
21
23
0
25 Jan 2023
Unifying Structure Reasoning and Language Model Pre-training for Complex Reasoning
Siyuan Wang
Zhongyu Wei
Jiarong Xu
Taishan Li
Zhihao Fan
LRM
36
5
0
21 Jan 2023
Coarse-to-Fine: Hierarchical Multi-task Learning for Natural Language Understanding
Zhaoye Fei
Yu Tian
Yongkang Wu
Xinyu Zhang
Yutao Zhu
...
Dejiang Kong
Ruofei Lai
Zhao Cao
Zhicheng Dou
Xipeng Qiu
51
1
0
19 Aug 2022
One Model, Multiple Modalities: A Sparsely Activated Approach for Text, Sound, Image, Video and Code
Yong Dai
Duyu Tang
Liangxin Liu
Minghuan Tan
Cong Zhou
Jingquan Wang
Zhangyin Feng
Fan Zhang
Xueyu Hu
Shuming Shi
VLM
MoE
21
26
0
12 May 2022
SkillNet-NLG: General-Purpose Natural Language Generation with a Sparsely Activated Approach
Junwei Liao
Duyu Tang
Fan Zhang
Shuming Shi
MoE
6
5
0
26 Apr 2022
Beyond Distillation: Task-level Mixture-of-Experts for Efficient Inference
Sneha Kudugunta
Yanping Huang
Ankur Bapna
M. Krikun
Dmitry Lepikhin
Minh-Thang Luong
Orhan Firat
MoE
119
106
0
24 Sep 2021
1