Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2212.07634
Cited By
Gradient-based Intra-attention Pruning on Pre-trained Language Models
15 December 2022
Ziqing Yang
Yiming Cui
Xin Yao
Shijin Wang
VLM
Re-assign community
ArXiv
PDF
HTML
Papers citing
"Gradient-based Intra-attention Pruning on Pre-trained Language Models"
3 / 3 papers shown
Title
COST-EFF: Collaborative Optimization of Spatial and Temporal Efficiency with Slenderized Multi-exit Language Models
Bowen Shen
Zheng Lin
Yuanxin Liu
Zhengxiao Liu
Lei Wang
Weiping Wang
VLM
36
4
0
27 Oct 2022
Extract then Distill: Efficient and Effective Task-Agnostic BERT Distillation
Cheng Chen
Yichun Yin
Lifeng Shang
Zhi Wang
Xin Jiang
Xiao Chen
Qun Liu
FedML
14
7
0
24 Apr 2021
GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding
Alex Jinpeng Wang
Amanpreet Singh
Julian Michael
Felix Hill
Omer Levy
Samuel R. Bowman
ELM
297
6,956
0
20 Apr 2018
1