Lightweight Deep Learning: An Overview

Ching Hao Wang, Kang Yang Huang, Yi Yao, Jun Cheng Chen, Hong Han Shuai, Wen Huang Cheng

Research output: Contribution to specialist publicationArticle

20 Scopus citations

Abstract

With the recent success of the deep neural networks (DNNs) in the field of artificial intelligence, the urge of deploying DNNs has drawn tremendous attention because it can benefit a wide range of applications on edge or embedded devices. Lightweight deep learning (DL) indicates the procedures of compressing DNN models into more compact ones, which are suitable to be executed on edge devices due to their limited resources and computational capabilities while maintaining comparable performance as the original. Currently, the approaches of model compression include but not limited to network pruning, quantization, knowledge distillation, neural architecture search. In this work, we present a fresh overview to summarize recent development and challenges for model compression.

Original languageEnglish
Pages1-12
Number of pages12
Specialist publicationIEEE Consumer Electronics Magazine
DOIs
StateAccepted/In press - 2022

Keywords

  • Computational modeling
  • Computer architecture
  • Costs
  • Deep learning
  • Neurons
  • Quantization (signal)
  • Training

Fingerprint

Dive into the research topics of 'Lightweight Deep Learning: An Overview'. Together they form a unique fingerprint.

Cite this