ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2203.03312
  4. Cited By
SkillNet-NLU: A Sparsely Activated Model for General-Purpose Natural
  Language Understanding

SkillNet-NLU: A Sparsely Activated Model for General-Purpose Natural Language Understanding

7 March 2022
Fan Zhang
Duyu Tang
Yong Dai
Cong Zhou
Shuangzhi Wu
Shuming Shi
    CLL
    MoE
ArXivPDFHTML

Papers citing "SkillNet-NLU: A Sparsely Activated Model for General-Purpose Natural Language Understanding"

10 / 10 papers shown
Title
ScaLearn: Simple and Highly Parameter-Efficient Task Transfer by
  Learning to Scale
ScaLearn: Simple and Highly Parameter-Efficient Task Transfer by Learning to Scale
Markus Frohmann
Carolin Holtermann
Shahed Masoudian
Anne Lauscher
Navid Rekabsaz
29
2
0
02 Oct 2023
LoraHub: Efficient Cross-Task Generalization via Dynamic LoRA
  Composition
LoraHub: Efficient Cross-Task Generalization via Dynamic LoRA Composition
Chengsong Huang
Qian Liu
Bill Yuchen Lin
Tianyu Pang
Chao Du
Min-Bin Lin
MoMe
36
182
0
25 Jul 2023
SkillNet-X: A Multilingual Multitask Model with Sparsely Activated
  Skills
SkillNet-X: A Multilingual Multitask Model with Sparsely Activated Skills
Zhangyin Feng
Yong Dai
Fan Zhang
Duyu Tang
Xiaocheng Feng
Shuangzhi Wu
Bing Qin
Yunbo Cao
Shuming Shi
MoE
29
0
0
28 Jun 2023
Modular Deep Learning
Modular Deep Learning
Jonas Pfeiffer
Sebastian Ruder
Ivan Vulić
E. Ponti
MoMe
OOD
21
73
0
22 Feb 2023
One Model for All Domains: Collaborative Domain-Prefix Tuning for
  Cross-Domain NER
One Model for All Domains: Collaborative Domain-Prefix Tuning for Cross-Domain NER
Xiang Chen
Lei Li
Q. Fei
Ningyu Zhang
Chuanqi Tan
Yong-jia Jiang
Fei Huang
Huajun Chen
21
23
0
25 Jan 2023
Unifying Structure Reasoning and Language Model Pre-training for Complex
  Reasoning
Unifying Structure Reasoning and Language Model Pre-training for Complex Reasoning
Siyuan Wang
Zhongyu Wei
Jiarong Xu
Taishan Li
Zhihao Fan
LRM
36
5
0
21 Jan 2023
Coarse-to-Fine: Hierarchical Multi-task Learning for Natural Language
  Understanding
Coarse-to-Fine: Hierarchical Multi-task Learning for Natural Language Understanding
Zhaoye Fei
Yu Tian
Yongkang Wu
Xinyu Zhang
Yutao Zhu
...
Dejiang Kong
Ruofei Lai
Zhao Cao
Zhicheng Dou
Xipeng Qiu
51
1
0
19 Aug 2022
One Model, Multiple Modalities: A Sparsely Activated Approach for Text,
  Sound, Image, Video and Code
One Model, Multiple Modalities: A Sparsely Activated Approach for Text, Sound, Image, Video and Code
Yong Dai
Duyu Tang
Liangxin Liu
Minghuan Tan
Cong Zhou
Jingquan Wang
Zhangyin Feng
Fan Zhang
Xueyu Hu
Shuming Shi
VLM
MoE
21
26
0
12 May 2022
SkillNet-NLG: General-Purpose Natural Language Generation with a
  Sparsely Activated Approach
SkillNet-NLG: General-Purpose Natural Language Generation with a Sparsely Activated Approach
Junwei Liao
Duyu Tang
Fan Zhang
Shuming Shi
MoE
6
5
0
26 Apr 2022
Beyond Distillation: Task-level Mixture-of-Experts for Efficient
  Inference
Beyond Distillation: Task-level Mixture-of-Experts for Efficient Inference
Sneha Kudugunta
Yanping Huang
Ankur Bapna
M. Krikun
Dmitry Lepikhin
Minh-Thang Luong
Orhan Firat
MoE
119
106
0
24 Sep 2021
1