ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2012.01266
  4. Cited By
Meta-KD: A Meta Knowledge Distillation Framework for Language Model
  Compression across Domains
v1v2 (latest)

Meta-KD: A Meta Knowledge Distillation Framework for Language Model Compression across Domains

Annual Meeting of the Association for Computational Linguistics (ACL), 2020
2 December 2020
Haojie Pan
Chengyu Wang
Minghui Qiu
Yichang Zhang
Yaliang Li
Yanjie Liang
ArXiv (abs)PDFHTML

Papers citing "Meta-KD: A Meta Knowledge Distillation Framework for Language Model Compression across Domains"

27 / 27 papers shown
Experts are all you need: A Composable Framework for Large Language Model Inference
Experts are all you need: A Composable Framework for Large Language Model Inference
S. Sridharan
Sourjya Roy
A. Raghunathan
Kaushik Roy
MoE
227
0
0
28 Nov 2025
PrunedLoRA: Robust Gradient-Based structured pruning for Low-rank Adaptation in Fine-tuning
PrunedLoRA: Robust Gradient-Based structured pruning for Low-rank Adaptation in Fine-tuning
Xin Yu
Cong Xie
Ziyu Zhao
Tiantian Fan
Lingzhou Xue
Zhi-Li Zhang
307
0
0
30 Sep 2025
SPADE: Structured Pruning and Adaptive Distillation for Efficient LLM-TTS
SPADE: Structured Pruning and Adaptive Distillation for Efficient LLM-TTS
T. Nguyen
Jaehun Kim
Ji-Hoon Kim
Shukjae Choi
Youshin Lim
Joon Son Chung
196
1
0
25 Sep 2025
DistilQwen2.5: Industrial Practices of Training Distilled Open Lightweight Language Models
DistilQwen2.5: Industrial Practices of Training Distilled Open Lightweight Language Models
Chengyu Wang
Junbing Yan
Yuanhao Yue
Yanjie Liang
359
6
0
21 Apr 2025
EvoP: Robust LLM Inference via Evolutionary Pruning
EvoP: Robust LLM Inference via Evolutionary Pruning
Shangyu Wu
Hongchao Du
Ying Xiong
Shuai Chen
Tei-Wei Kuo
Nan Guan
Chun Jason Xue
701
3
0
19 Feb 2025
A Hybrid Cross-Stage Coordination Pre-ranking Model for Online Recommendation Systems
A Hybrid Cross-Stage Coordination Pre-ranking Model for Online Recommendation SystemsThe Web Conference (WWW), 2025
Binglei Zhao
Houying Qi
Guang Xu
Mian Ma
Xiwei Zhao
Feng Mei
Sulong Xu
Jinghe Hu
297
5
0
17 Feb 2025
MoDeGPT: Modular Decomposition for Large Language Model Compression
MoDeGPT: Modular Decomposition for Large Language Model CompressionInternational Conference on Learning Representations (ICLR), 2024
Chi-Heng Lin
Shangqian Gao
James Seale Smith
Abhishek Patel
Shikhar Tuli
Yilin Shen
Hongxia Jin
Yen-Chang Hsu
894
35
0
19 Aug 2024
MIDAS: Multi-level Intent, Domain, And Slot Knowledge Distillation for Multi-turn NLU
MIDAS: Multi-level Intent, Domain, And Slot Knowledge Distillation for Multi-turn NLUNorth American Chapter of the Association for Computational Linguistics (NAACL), 2024
Yan Li
So-Eon Kim
Seong-Bae Park
S. Han
498
2
0
15 Aug 2024
Weight Copy and Low-Rank Adaptation for Few-Shot Distillation of Vision
  Transformers
Weight Copy and Low-Rank Adaptation for Few-Shot Distillation of Vision Transformers
Diana-Nicoleta Grigore
Mariana-Iuliana Georgescu
J. A. Justo
T. Johansen
Andreea-Iuliana Ionescu
Radu Tudor Ionescu
381
2
0
14 Apr 2024
CATS: Contextually-Aware Thresholding for Sparsity in Large Language
  Models
CATS: Contextually-Aware Thresholding for Sparsity in Large Language Models
Je-Yong Lee
Donghyun Lee
Genghan Zhang
Mo Tiwari
Azalia Mirhoseini
337
37
0
12 Apr 2024
Hierarchical Skip Decoding for Efficient Autoregressive Text Generation
Hierarchical Skip Decoding for Efficient Autoregressive Text Generation
Yunqi Zhu
Xuebing Yang
Yuanyuan Wu
Wensheng Zhang
400
6
0
22 Mar 2024
CLLMs: Consistency Large Language Models
CLLMs: Consistency Large Language Models
Siqi Kou
Lanxiang Hu
Zhe He
Zhijie Deng
Hao Zhang
512
61
0
28 Feb 2024
Model Compression and Efficient Inference for Large Language Models: A
  Survey
Model Compression and Efficient Inference for Large Language Models: A Survey
Wenxiao Wang
Wei Chen
Yicong Luo
Yongliu Long
Zhengkai Lin
Liye Zhang
Binbin Lin
Deng Cai
Xiaofei He
MQ
362
93
0
15 Feb 2024
One-Shot Sensitivity-Aware Mixed Sparsity Pruning for Large Language
  Models
One-Shot Sensitivity-Aware Mixed Sparsity Pruning for Large Language Models
Hang Shao
Bei Liu
Bo Xiao
Ke Zeng
Guanglu Wan
Yanmin Qian
319
30
0
14 Oct 2023
Position: Key Claims in LLM Research Have a Long Tail of Footnotes
Position: Key Claims in LLM Research Have a Long Tail of FootnotesInternational Conference on Machine Learning (ICML), 2023
Anna Rogers
A. Luccioni
547
24
0
14 Aug 2023
Are Intermediate Layers and Labels Really Necessary? A General Language
  Model Distillation Method
Are Intermediate Layers and Labels Really Necessary? A General Language Model Distillation MethodAnnual Meeting of the Association for Computational Linguistics (ACL), 2023
Shicheng Tan
Weng Lam Tam
Yuanchun Wang
Wenwen Gong
Shuo Zhao
Peng Zhang
Jie Tang
VLM
182
1
0
11 Jun 2023
Domain Private Transformers for Multi-Domain Dialog Systems
Domain Private Transformers for Multi-Domain Dialog SystemsConference on Empirical Methods in Natural Language Processing (EMNLP), 2023
Anmol Kabra
Ethan R. Elenberg
266
0
0
23 May 2023
LLM-Pruner: On the Structural Pruning of Large Language Models
LLM-Pruner: On the Structural Pruning of Large Language ModelsNeural Information Processing Systems (NeurIPS), 2023
Xinyin Ma
Gongfan Fang
Xinchao Wang
880
766
0
19 May 2023
Few-Shot Learning of Compact Models via Task-Specific Meta Distillation
Few-Shot Learning of Compact Models via Task-Specific Meta DistillationIEEE Workshop/Winter Conference on Applications of Computer Vision (WACV), 2022
Yong Wu
Shekhor Chanda
M. Hosseinzadeh
Zhi Liu
Yang Wang
VLM
317
10
0
18 Oct 2022
Meta Learning for Natural Language Processing: A Survey
Meta Learning for Natural Language Processing: A SurveyNorth American Chapter of the Association for Computational Linguistics (NAACL), 2022
Hung-yi Lee
Shang-Wen Li
Ngoc Thang Vu
441
55
0
03 May 2022
EasyNLP: A Comprehensive and Easy-to-use Toolkit for Natural Language
  Processing
EasyNLP: A Comprehensive and Easy-to-use Toolkit for Natural Language ProcessingConference on Empirical Methods in Natural Language Processing (EMNLP), 2022
Chengyu Wang
Minghui Qiu
Chen Shi
Taolin Zhang
Tingting Liu
Lei Li
Jiadong Wang
Ming Wang
Yanjie Liang
W. Lin
273
22
0
30 Apr 2022
DistilCSE: Effective Knowledge Distillation For Contrastive Sentence
  Embeddings
DistilCSE: Effective Knowledge Distillation For Contrastive Sentence Embeddings
Chaochen Gao
Xing Wu
Peng Wang
Jue Wang
Liangjun Zang
Zhongyuan Wang
Songlin Hu
204
6
0
10 Dec 2021
Edge-Cloud Polarization and Collaboration: A Comprehensive Survey for AI
Edge-Cloud Polarization and Collaboration: A Comprehensive Survey for AIIEEE Transactions on Knowledge and Data Engineering (TKDE), 2021
Jiangchao Yao
Shengyu Zhang
Yang Yao
Feng Wang
Jianxin Ma
...
Kun Kuang
Chao-Xiang Wu
Leilei Gan
Jingren Zhou
Hongxia Yang
441
155
0
11 Nov 2021
HRKD: Hierarchical Relational Knowledge Distillation for Cross-domain
  Language Model Compression
HRKD: Hierarchical Relational Knowledge Distillation for Cross-domain Language Model Compression
Chenhe Dong
Yaliang Li
Ying Shen
Minghui Qiu
VLM
371
8
0
16 Oct 2021
Learning to Teach with Student Feedback
Learning to Teach with Student Feedback
Yitao Liu
Tianxiang Sun
Xipeng Qiu
Xuanjing Huang
VLM
192
6
0
10 Sep 2021
BERT Learns to Teach: Knowledge Distillation with Meta Learning
BERT Learns to Teach: Knowledge Distillation with Meta LearningAnnual Meeting of the Association for Computational Linguistics (ACL), 2021
Wangchunshu Zhou
Canwen Xu
Julian McAuley
378
110
0
08 Jun 2021
EasyTransfer -- A Simple and Scalable Deep Transfer Learning Platform
  for NLP Applications
EasyTransfer -- A Simple and Scalable Deep Transfer Learning Platform for NLP ApplicationsInternational Conference on Information and Knowledge Management (CIKM), 2020
Minghui Qiu
Peng Li
Chengyu Wang
Hanjie Pan
Yaliang Li
...
Jun Yang
Yaliang Li
Yanjie Liang
Deng Cai
Jialin Li
VLMSyDa
414
20
0
18 Nov 2020
1
Page 1 of 1