Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2109.06243
Cited By
KroneckerBERT: Learning Kronecker Decomposition for Pre-trained Language Models via Knowledge Distillation
13 September 2021
Marzieh S. Tahaei
Ella Charlaix
V. Nia
A. Ghodsi
Mehdi Rezagholizadeh
Re-assign community
ArXiv
PDF
HTML
Papers citing
"KroneckerBERT: Learning Kronecker Decomposition for Pre-trained Language Models via Knowledge Distillation"
7 / 7 papers shown
Title
TQCompressor: improving tensor decomposition methods in neural networks via permutations
V. Abronin
A. Naumov
D. Mazur
D. Bystrov
K. Tsarova
Ar. Melnikov
Ivan V. Oseledets
S. Dolgov
R. Brasher
M. Perelshtein
18
6
0
29 Jan 2024
LoSparse: Structured Compression of Large Language Models based on Low-Rank and Sparse Approximation
Yixiao Li
Yifan Yu
Qingru Zhang
Chen Liang
Pengcheng He
Weizhu Chen
Tuo Zhao
17
64
0
20 Jun 2023
Parameter-efficient Model Adaptation for Vision Transformers
Xuehai He
Chunyuan Li
Pengchuan Zhang
Jianwei Yang
X. Wang
20
80
0
29 Mar 2022
Kronecker Decomposition for GPT Compression
Ali Edalati
Marzieh S. Tahaei
Ahmad Rashid
V. Nia
J. Clark
Mehdi Rezagholizadeh
26
33
0
15 Oct 2021
BERT-of-Theseus: Compressing BERT by Progressive Module Replacing
Canwen Xu
Wangchunshu Zhou
Tao Ge
Furu Wei
Ming Zhou
210
196
0
07 Feb 2020
Megatron-LM: Training Multi-Billion Parameter Language Models Using Model Parallelism
M. Shoeybi
M. Patwary
Raul Puri
P. LeGresley
Jared Casper
Bryan Catanzaro
MoE
243
1,791
0
17 Sep 2019
GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding
Alex Jinpeng Wang
Amanpreet Singh
Julian Michael
Felix Hill
Omer Levy
Samuel R. Bowman
ELM
294
6,927
0
20 Apr 2018
1