ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2311.13811
56
0
v1v2v3 (latest)

Education distillation:getting student models to learn in shcools

23 November 2023
Ling Feng
Danyang Li
Tianhao Wu
Xuliang Duan
    FedML
ArXiv (abs)PDFHTML
Abstract

Knowledge distillation is one of the methods for model compression, and existing knowledge distillation techniques focus on how to improve the distillation algorithm so as to enhance the distillation efficdiency. This paper introduces dynamic incremental learning into knowledge distillation and proposes a distillation strategy for education distillation. Specifically, it is proposed to look at fragmented student models divided from the full student model as low models. As the grade level rises, fragmented student models deepen in conjunction with designed teaching reference layers, while learning and distilling from more teacher models. By moving from lower to higher grades, fragmented student models were gradually integrated into a complete target student model, and the performance of the student models gradually improved from lower to senior grades of the stage. Education distillation strategies combined with distillation algorithms outperform the results of single distillation algorithms on the public dataset CIFAR100,Caltech256, Food-101 dataset.

View on arXiv
@article{feng2025_2311.13811,
  title={ Education distillation:getting student models to learn in shcools },
  author={ Ling Feng and Tianhao Wu and Xiangrong Ren and Zhi Jing and Xuliang Duan },
  journal={arXiv preprint arXiv:2311.13811},
  year={ 2025 }
}
Comments on this paper