Lightweight Deep Learning: An Overview

Ching Hao Wang*, Kang Yang Huang, Yi Yao, Jun Cheng Chen, Hong Han Shuai, Wen Huang Cheng

*Corresponding author for this work

Research output: Contribution to specialist publicationArticle

31 Scopus citations

Abstract

With the recent success of the deep neural networks (DNNs) in the field of artificial intelligence, the urge of deploying DNNs has drawn tremendous attention because it can benefit a wide range of applications on edge or embedded devices. Lightweight deep learning indicates the procedures of compressing DNN models into more compact ones, which are suitable to be executed on edge devices due to their limited resources and computational capabilities while maintaining comparable performance to the original. Currently, the approaches of model compression include but are not limited to network pruning, quantization, knowledge distillation, and neural architecture search. In this work, we present a fresh overview to summarize recent development and challenges for model compression.

Original languageEnglish
Pages51-64
Number of pages14
Volume13
No4
Specialist publicationIEEE Consumer Electronics Magazine
DOIs
StatePublished - 1 Jul 2024

Fingerprint

Dive into the research topics of 'Lightweight Deep Learning: An Overview'. Together they form a unique fingerprint.

Cite this