Memory-Replay Knowledge Distillation

Knowledge Distillation (KD), which transfers the knowledge from a teacher to a student network by penalizing their Kullback–Leibler (KL) divergence, is a widely used tool for Deep Neural Network (DNN) compression in intelligent sensor systems. Traditional KD uses pre-trained teacher, while self-KD d...

Full description

Bibliographic Details
Main Authors: Jiyue Wang, Pei Zhang, Yanxiong Li
Format: Article
Language:English
Published: MDPI AG 2021-04-01
Series:Sensors
Subjects:
Online Access:https://www.mdpi.com/1424-8220/21/8/2792