Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2109.08359
Cited By
Distilling Linguistic Context for Language Model Compression
17 September 2021
Geondo Park
Gyeongman Kim
Eunho Yang
Re-assign community
ArXiv
PDF
HTML
Papers citing
"Distilling Linguistic Context for Language Model Compression"
4 / 4 papers shown
Title
Multi-Level Optimal Transport for Universal Cross-Tokenizer Knowledge Distillation on Language Models
Xiao Cui
Mo Zhu
Yulei Qin
Liang Xie
Wengang Zhou
H. Li
81
2
0
19 Dec 2024
Revisiting Intermediate Layer Distillation for Compressing Language Models: An Overfitting Perspective
Jongwoo Ko
Seungjoon Park
Minchan Jeong
S. Hong
Euijai Ahn
Duhyeuk Chang
Se-Young Yun
19
6
0
03 Feb 2023
BERT-of-Theseus: Compressing BERT by Progressive Module Replacing
Canwen Xu
Wangchunshu Zhou
Tao Ge
Furu Wei
Ming Zhou
213
196
0
07 Feb 2020
GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding
Alex Jinpeng Wang
Amanpreet Singh
Julian Michael
Felix Hill
Omer Levy
Samuel R. Bowman
ELM
294
6,927
0
20 Apr 2018
1