Multi-level Distillation of Semantic Knowledge for Pre-training
Multilingual Language ModelConference on Empirical Methods in Natural Language Processing (EMNLP), 2022 |
Contrastive Representation DistillationInternational Conference on Learning Representations (ICLR), 2019 |