Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2307.08436
Cited By
DOT: A Distillation-Oriented Trainer
17 July 2023
Borui Zhao
Quan Cui
Renjie Song
Jiajun Liang
Re-assign community
ArXiv
PDF
HTML
Papers citing
"DOT: A Distillation-Oriented Trainer"
3 / 3 papers shown
Title
Knowledge Distillation in Wide Neural Networks: Risk Bound, Data Efficiency and Imperfect Teacher
Guangda Ji
Zhanxing Zhu
51
42
0
20 Oct 2020
On Large-Batch Training for Deep Learning: Generalization Gap and Sharp Minima
N. Keskar
Dheevatsa Mudigere
J. Nocedal
M. Smelyanskiy
P. T. P. Tang
ODL
273
2,888
0
15 Sep 2016
ImageNet Large Scale Visual Recognition Challenge
Olga Russakovsky
Jia Deng
Hao Su
J. Krause
S. Satheesh
...
A. Karpathy
A. Khosla
Michael S. Bernstein
Alexander C. Berg
Li Fei-Fei
VLM
ObjD
287
39,194
0
01 Sep 2014
1