Skip to content

Latest commit

 

History

History
11 lines (7 loc) · 320 Bytes

200209 Understanding and Improving Knowledge Distillation.md

File metadata and controls

11 lines (7 loc) · 320 Bytes

https://arxiv.org/abs/2002.03532

Understanding and Improving Knowledge Distillation (Jiaxi Tang, Rakesh Shivanna, Zhe Zhao, Dong Lin, Anima Singh, Ed H. Chi, Sagar Jain)

KD의 효과를 분석.

  1. label smoothing
  2. teacher의 confidence를 활용해 example를 reweighting
  3. logit에 prior를 제공

#distillation