ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2003.03622
  4. Cited By
Explaining Knowledge Distillation by Quantifying the Knowledge

Explaining Knowledge Distillation by Quantifying the Knowledge

Computer Vision and Pattern Recognition (CVPR), 2025
7 March 2020
Xu Cheng
Zhefan Rao
Yilan Chen
Quanshi Zhang
ArXiv (abs)PDFHTML

Papers citing "Explaining Knowledge Distillation by Quantifying the Knowledge"

50 / 60 papers shown
Title
TopKD: Top-scaled Knowledge Distillation
TopKD: Top-scaled Knowledge Distillation
Qi Wang
Jinjia Zhou
23
0
0
06 Aug 2025
A Transformer-in-Transformer Network Utilizing Knowledge Distillation for Image Recognition
A Transformer-in-Transformer Network Utilizing Knowledge Distillation for Image Recognition
Dewan Tauhid Rahman
Yeahia Sarker
Antar Mazumder
Md. Shamim Anower
ViT
101
0
0
24 Feb 2025
Decoupling Dark Knowledge via Block-wise Logit Distillation for
  Feature-level Alignment
Decoupling Dark Knowledge via Block-wise Logit Distillation for Feature-level Alignment
Chengting Yu
Fengzhao Zhang
Ruizhe Chen
Zuozhu Liu
Shurun Tan
Er-ping Li
Aili Wang
155
3
0
03 Nov 2024
Multi-Level Feature Distillation of Joint Teachers Trained on Distinct
  Image Datasets
Multi-Level Feature Distillation of Joint Teachers Trained on Distinct Image Datasets
Adrian Iordache
B. Alexe
Radu Tudor Ionescu
163
2
0
29 Oct 2024
InFiConD: Interactive No-code Fine-tuning with Concept-based Knowledge
  Distillation
InFiConD: Interactive No-code Fine-tuning with Concept-based Knowledge Distillation
Jinbin Huang
Wenbin He
Liang Gou
Liu Ren
Chris Bryan
174
0
0
25 Jun 2024
Exploring Dark Knowledge under Various Teacher Capacities and Addressing Capacity Mismatch
Exploring Dark Knowledge under Various Teacher Capacities and Addressing Capacity Mismatch
Wen-Shu Fan
Xin-Chun Li
Bowen Tao
126
2
0
21 May 2024
CKD: Contrastive Knowledge Distillation from A Sample-wise Perspective
CKD: Contrastive Knowledge Distillation from A Sample-wise Perspective
Wencheng Zhu
Xin Zhou
Pengfei Zhu
Yu Wang
Qinghua Hu
VLM
220
2
0
22 Apr 2024
Weight Copy and Low-Rank Adaptation for Few-Shot Distillation of Vision
  Transformers
Weight Copy and Low-Rank Adaptation for Few-Shot Distillation of Vision Transformers
Diana-Nicoleta Grigore
Mariana-Iuliana Georgescu
J. A. Justo
T. Johansen
Andreea-Iuliana Ionescu
Radu Tudor Ionescu
137
1
0
14 Apr 2024
Enhancing Metaphor Detection through Soft Labels and Target Word
  Prediction
Enhancing Metaphor Detection through Soft Labels and Target Word Prediction
Kaidi Jia
Rongsheng Li
76
0
0
27 Mar 2024
Enabling Generalized Zero-shot Learning Towards Unseen Domains by Intrinsic Learning from Redundant LLM Semantics
Enabling Generalized Zero-shot Learning Towards Unseen Domains by Intrinsic Learning from Redundant LLM Semantics
Jiaqi Yue
Jiancheng Zhao
Chunhui Zhao
Biao Huang
110
0
0
21 Mar 2024
Two-Stage Multi-task Self-Supervised Learning for Medical Image
  Segmentation
Two-Stage Multi-task Self-Supervised Learning for Medical Image Segmentation
Binyan Hu
•. A. K. Qin
SSL
71
0
0
11 Feb 2024
Iterative Data Smoothing: Mitigating Reward Overfitting and
  Overoptimization in RLHF
Iterative Data Smoothing: Mitigating Reward Overfitting and Overoptimization in RLHF
Banghua Zhu
Michael I. Jordan
Jiantao Jiao
120
40
0
29 Jan 2024
Zone Evaluation: Revealing Spatial Bias in Object Detection
Zone Evaluation: Revealing Spatial Bias in Object Detection
Zhaohui Zheng
Yuming Chen
Qibin Hou
Xiang Li
Ping Wang
Ming-Ming Cheng
ObjD
140
6
0
20 Oct 2023
Towards the Fundamental Limits of Knowledge Transfer over Finite Domains
Towards the Fundamental Limits of Knowledge Transfer over Finite Domains
Qingyue Zhao
Banghua Zhu
168
4
0
11 Oct 2023
Computation-efficient Deep Learning for Computer Vision: A Survey
Computation-efficient Deep Learning for Computer Vision: A Survey
Yulin Wang
Yizeng Han
Chaofei Wang
Shiji Song
Qi Tian
Gao Huang
VLM
186
26
0
27 Aug 2023
Dissecting RGB-D Learning for Improved Multi-modal Fusion
Dissecting RGB-D Learning for Improved Multi-modal Fusion
Hao Chen
Hao Zhou
Yunshu Zhang
Zheng Lin
Yongjian Deng
203
1
0
19 Aug 2023
Teacher-Student Architecture for Knowledge Distillation: A Survey
Teacher-Student Architecture for Knowledge Distillation: A Survey
Chengming Hu
Xuan Li
Danyang Liu
Haolun Wu
Xi Chen
Ju Wang
Xue Liu
140
26
0
08 Aug 2023
Towards Better Query Classification with Multi-Expert Knowledge
  Condensation in JD Ads Search
Towards Better Query Classification with Multi-Expert Knowledge Condensation in JD Ads Search
Hai-Jian Ke
Ming Pang
Zheng Fang
Xue Jiang
Xi-Wei Zhao
Changping Peng
Zhangang Lin
Jinghe Hu
Jingping Shao
177
0
0
02 Aug 2023
Frameless Graph Knowledge Distillation
Frameless Graph Knowledge Distillation
Dai Shi
Zhiqi Shao
Yi Guo
Junbin Gao
92
4
0
13 Jul 2023
Exploring the Lottery Ticket Hypothesis with Explainability Methods:
  Insights into Sparse Network Performance
Exploring the Lottery Ticket Hypothesis with Explainability Methods: Insights into Sparse Network Performance
Shantanu Ghosh
Kayhan Batmanghelich
103
1
0
07 Jul 2023
Dividing and Conquering a BlackBox to a Mixture of Interpretable Models:
  Route, Interpret, Repeat
Dividing and Conquering a BlackBox to a Mixture of Interpretable Models: Route, Interpret, Repeat
Shantanu Ghosh
K. Yu
Forough Arabshahi
Kayhan Batmanghelich
MoE
128
14
0
07 Jul 2023
Leveraging Synthetic Targets for Machine Translation
Leveraging Synthetic Targets for Machine Translation
Sarthak Mittal
Oleksii Hrinchuk
Oleksii Kuchaiev
82
2
0
07 May 2023
Improving Knowledge Distillation via Transferring Learning Ability
Improving Knowledge Distillation via Transferring Learning Ability
Long Liu
Tong Li
Hui Cheng
72
1
0
24 Apr 2023
Towards a Smaller Student: Capacity Dynamic Distillation for Efficient
  Image Retrieval
Towards a Smaller Student: Capacity Dynamic Distillation for Efficient Image Retrieval
Yi Xie
Huaidong Zhang
Xuemiao Xu
Jianqing Zhu
Shengfeng He
VLM
139
16
0
16 Mar 2023
Graph-based Knowledge Distillation: A survey and experimental evaluation
Graph-based Knowledge Distillation: A survey and experimental evaluation
Jing Liu
Tongya Zheng
Guanzheng Zhang
Qinfen Hao
113
10
0
27 Feb 2023
Practical Knowledge Distillation: Using DNNs to Beat DNNs
Practical Knowledge Distillation: Using DNNs to Beat DNNs
Chungman Lee
Pavlos Anastasios Apostolopulos
Igor L. Markov
FedML
102
1
0
23 Feb 2023
Understanding Self-Distillation in the Presence of Label Noise
Understanding Self-Distillation in the Presence of Label Noise
Rudrajit Das
Sujay Sanghavi
151
19
0
30 Jan 2023
Teacher-Student Architecture for Knowledge Learning: A Survey
Teacher-Student Architecture for Knowledge Learning: A Survey
Chengming Hu
Xuan Li
Dan Liu
Xi Chen
Ju Wang
Xue Liu
158
39
0
28 Oct 2022
On effects of Knowledge Distillation on Transfer Learning
On effects of Knowledge Distillation on Transfer Learning
Sushil Thapa
64
2
0
18 Oct 2022
Asymmetric Temperature Scaling Makes Larger Networks Teach Well Again
Asymmetric Temperature Scaling Makes Larger Networks Teach Well Again
Xin-Chun Li
Wenxuan Fan
Shaoming Song
Yinchuan Li
Bingshuai Li
Yunfeng Shao
De-Chuan Zhan
168
32
0
10 Oct 2022
Quantifying the Knowledge in a DNN to Explain Knowledge Distillation for
  Classification
Quantifying the Knowledge in a DNN to Explain Knowledge Distillation for Classification
Quanshi Zhang
Xu Cheng
Yilan Chen
Zhefan Rao
97
38
0
18 Aug 2022
Explaining Deepfake Detection by Analysing Image Matching
Explaining Deepfake Detection by Analysing Image Matching
S. Dong
Jin Wang
Jiajun Liang
Haoqiang Fan
Renhe Ji
96
58
0
20 Jul 2022
Narrowing the Coordinate-frame Gap in Behavior Prediction Models:
  Distillation for Efficient and Accurate Scene-centric Motion Forecasting
Narrowing the Coordinate-frame Gap in Behavior Prediction Models: Distillation for Efficient and Accurate Scene-centric Motion Forecasting
DiJia Su
B. Douillard
Rami Al-Rfou
C. Park
Benjamin Sapp
115
10
0
08 Jun 2022
A General Multiple Data Augmentation Based Framework for Training Deep
  Neural Networks
A General Multiple Data Augmentation Based Framework for Training Deep Neural Networks
Bin Hu
Yu Sun
•. A. K. Qin
AI4CE
93
0
0
29 May 2022
Heterogeneous Collaborative Learning for Personalized Healthcare
  Analytics via Messenger Distillation
Heterogeneous Collaborative Learning for Personalized Healthcare Analytics via Messenger Distillation
Guanhua Ye
Tong Chen
Yawen Li
Li-zhen Cui
Quoc Viet Hung Nguyen
Hongzhi Yin
150
9
0
27 May 2022
Compressing Deep Graph Neural Networks via Adversarial Knowledge
  Distillation
Compressing Deep Graph Neural Networks via Adversarial Knowledge Distillation
Huarui He
Jie Wang
Zhanqiu Zhang
Feng Wu
106
46
0
24 May 2022
Towards Feature Distribution Alignment and Diversity Enhancement for
  Data-Free Quantization
Towards Feature Distribution Alignment and Diversity Enhancement for Data-Free Quantization
Yangcheng Gao
Zhao Zhang
Richang Hong
Haijun Zhang
Jicong Fan
Shuicheng Yan
MQ
101
10
0
30 Apr 2022
Decoupled Knowledge Distillation
Decoupled Knowledge Distillation
Borui Zhao
Quan Cui
Renjie Song
Yiyu Qiu
Jiajun Liang
194
626
0
16 Mar 2022
From Anecdotal Evidence to Quantitative Evaluation Methods: A Systematic
  Review on Evaluating Explainable AI
From Anecdotal Evidence to Quantitative Evaluation Methods: A Systematic Review on Evaluating Explainable AI
Meike Nauta
Jan Trienes
Shreyasi Pathak
Elisa Nguyen
Michelle Peters
Yasmin Schmitt
Jorg Schlotterer
M. V. Keulen
C. Seifert
ELMXAI
275
478
0
20 Jan 2022
Class-Incremental Continual Learning into the eXtended DER-verse
Class-Incremental Continual Learning into the eXtended DER-verse
Matteo Boschini
Lorenzo Bonicelli
Pietro Buzzega
Angelo Porrello
Simone Calderara
CLLBDL
137
153
0
03 Jan 2022
Boosting Unsupervised Domain Adaptation with Soft Pseudo-label and
  Curriculum Learning
Boosting Unsupervised Domain Adaptation with Soft Pseudo-label and Curriculum Learning
Shengjia Zhang
Tiancheng Lin
Yi Tian Xu
142
5
0
03 Dec 2021
Semi-supervised Domain Adaptation via Sample-to-Sample Self-Distillation
Semi-supervised Domain Adaptation via Sample-to-Sample Self-Distillation
Jeongbeen Yoon
Dahyun Kang
Minsu Cho
TTA
107
44
0
29 Nov 2021
Visualizing the Emergence of Intermediate Visual Patterns in DNNs
Visualizing the Emergence of Intermediate Visual Patterns in DNNs
Mingjie Li
Shaobo Wang
Quanshi Zhang
139
11
0
05 Nov 2021
Instance-Conditional Knowledge Distillation for Object Detection
Instance-Conditional Knowledge Distillation for Object Detection
Zijian Kang
Peizhen Zhang
Xinming Zhang
Jian Sun
N. Zheng
145
80
0
25 Oct 2021
Visualizing the embedding space to explain the effect of knowledge
  distillation
Visualizing the embedding space to explain the effect of knowledge distillation
Hyun Seung Lee
C. Wallraven
103
1
0
09 Oct 2021
Partial to Whole Knowledge Distillation: Progressive Distilling
  Decomposed Knowledge Boosts Student Better
Partial to Whole Knowledge Distillation: Progressive Distilling Decomposed Knowledge Boosts Student Better
Xuanyang Zhang
Xinming Zhang
Jian Sun
91
1
0
26 Sep 2021
Efficient Action Recognition Using Confidence Distillation
Efficient Action Recognition Using Confidence Distillation
Shervin Manzuri Shalmani
Fei Chiang
Ronghuo Zheng
174
6
0
05 Sep 2021
Embracing the Dark Knowledge: Domain Generalization Using Regularized
  Knowledge Distillation
Embracing the Dark Knowledge: Domain Generalization Using Regularized Knowledge Distillation
Yufei Wang
Haoliang Li
Lap-pui Chau
Alex C. Kot
FedML
90
47
0
06 Jul 2021
Dynamic Knowledge Distillation With Noise Elimination for RGB-D Salient
  Object Detection
Dynamic Knowledge Distillation With Noise Elimination for RGB-D Salient Object Detection
Guangyu Ren
Yinxiao Yu
Hengyan Liu
Tania Stathaki
126
7
0
17 Jun 2021
Knowledge Distillation as Semiparametric Inference
Knowledge Distillation as Semiparametric Inference
Tri Dao
G. Kamath
Vasilis Syrgkanis
Lester W. Mackey
110
32
0
20 Apr 2021
12
Next