Understanding the Gains from Repeated Self-Distillation

Neural Information Processing Systems 

Self-Distillation is a special type of knowledge distillation where the student model has the same architecture as the teacher model.

Similar Docs  Excel Report  more

TitleSimilaritySource
None found