The Role of Teacher Calibration in Knowledge Distillation

Knowledge Distillation (KD) has emerged as an effective model compression technique in deep learning, enabling the transfer of knowledge from a large teacher model to a compact student model. While KD has demonstrated significant success, it is not yet fully understood which factors contribute to im...

Full description

Saved in:
Bibliographic Details
Main Authors: Suyoung Kim, Seonguk Park, Junhoo Lee, Nojun Kwak
Format: Article
Language:English
Published: IEEE 2025-01-01
Series:IEEE Access
Subjects:
Online Access:https://ieeexplore.ieee.org/document/11062864/
Tags: Add Tag
No Tags, Be the first to tag this record!