Knowledge distillation in deep learning and its applications

Abdolmaged Alkhulaifi, Fahad Alsahli, Irfan Ahmad*

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

82 Scopus citations

Abstract

Deep learning based models are relatively large, and it is hard to deploy such models on resource-limited devices such as mobile phones and embedded devices. One possible solution is knowledge distillation whereby a smaller model (student model) is trained by utilizing the information from a larger model (teacher model). In this paper, we present an outlook of knowledge distillation techniques applied to deep learning models. To compare the performances of different techniques, we propose a new metric called distillation metric which compares different knowledge distillation solutions based on models' sizes and accuracy scores. Based on the survey, some interesting conclusions are drawn and presented in this paper including the current challenges and possible research directions.

Original languageEnglish
Pages (from-to)1-24
Number of pages24
JournalPeerJ Computer Science
Volume7
DOIs
StatePublished - Apr 2021

Bibliographical note

Publisher Copyright:
© 2021 Alkhulaifi et al. All Rights Reserved.

Keywords

  • Deep learning
  • Knowledge distillation
  • Model compression
  • Student model
  • Teacher model
  • Transferring knowledge

ASJC Scopus subject areas

  • General Computer Science

Fingerprint

Dive into the research topics of 'Knowledge distillation in deep learning and its applications'. Together they form a unique fingerprint.

Cite this