ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2110.15946
  4. Cited By
Estimating and Maximizing Mutual Information for Knowledge Distillation

Estimating and Maximizing Mutual Information for Knowledge Distillation

29 October 2021
A. Shrivastava
Yanjun Qi
Vicente Ordonez
ArXivPDFHTML

Papers citing "Estimating and Maximizing Mutual Information for Knowledge Distillation"

2 / 2 papers shown
Title
Teacher-Student Architecture for Knowledge Distillation: A Survey
Teacher-Student Architecture for Knowledge Distillation: A Survey
Chengming Hu
Xuan Li
Danyang Liu
Haolun Wu
Xi Chen
Ju Wang
Xue Liu
21
16
0
08 Aug 2023
Double-Hard Debias: Tailoring Word Embeddings for Gender Bias Mitigation
Double-Hard Debias: Tailoring Word Embeddings for Gender Bias Mitigation
Tianlu Wang
Xi Victoria Lin
Nazneen Rajani
Bryan McCann
Vicente Ordonez
Caimng Xiong
CVBM
151
54
0
03 May 2020
1