Wasserstein Distance Rivals Kullback-Leibler Divergence for Knowledge Distillation Jiaming Lv, Haoyuan Y ang
–Neural Information Processing Systems
Since pioneering work of Hinton et al., knowledge distillation based on Kullback-Leibler Divergence (KL-Div) has been predominant, and recently its variants have
Neural Information Processing Systems
Oct-10-2025, 06:39:36 GMT
- Country:
- Asia > China
- Guangxi Province > Nanning (0.04)
- Liaoning Province > Dalian (0.04)
- North America > United States
- California > Santa Clara County > Palo Alto (0.04)
- Asia > China
- Genre:
- Research Report > Experimental Study (0.93)
- Industry:
- Education (0.69)
- Technology:
- Information Technology
- Artificial Intelligence
- Machine Learning
- Neural Networks > Deep Learning (0.93)
- Statistical Learning (1.00)
- Natural Language (1.00)
- Representation & Reasoning (1.00)
- Vision (1.00)
- Machine Learning
- Data Science > Data Mining (0.67)
- Artificial Intelligence
- Information Technology