Tuesday Aug 08, 2023
ICLR 2023 - Towards Understanding Ensemble, Knowledge Distillation and Self-Distillation in Deep Learning
In this episode we discuss Towards Understanding Ensemble, Knowledge Distillation and Self-Distillation in Deep Learning by Zeyuan Allen-Zhu, Yuanzhi Li. The paper explores how ensembles of deep learning models can improve test accuracy and be distilled into a single model using knowledge distillation. It presents a theoretical framework that shows how ensembles can enhance test accuracy when the data has a multi-view structure. The paper also highlights the presence of "dark knowledge" within ensemble outputs and demonstrates that self-distillation combines ensemble learning and knowledge distillation for improved test accuracy.
Comments (0)
To leave or reply to comments, please download free Podbean or
No Comments
To leave or reply to comments,
please download free Podbean App.