@inproceedings{a575117d5573403986d2eba919799631,
title = "Supporting compressed-sparse activations and weights on SIMD-like accelerator for sparse convolutional neural networks",
abstract = "Sparsity is widely observed in convolutional neural networks by zeroing a large portion of both activations and weights without impairing the result. By keeping the data in a compressed-sparse format, the energy consumption could be considerably cut down due to less memory traffic. However, the wide SIMD-like MAC engine adopted in many CNN accelerators can not support the compressed input due to the data misalignment. In this work, a novel Dual Indexing Module (DIM) is proposed to efficiently handle the alignment issue where activations and weights are both kept in compressed-sparse format. The DIM is implemented in a representative SIMD-like CNN accelerator, and able to exploit both compressed-sparse activations and weights. The synthesis results with 40nm technology have shown that DIM can enhance up to 46% of energy consumption and 55.4% Energy-Delay-Product (EDP).",
author = "Lin, {Chien Yu} and Bo-Cheng Lai",
year = "2018",
month = feb,
day = "20",
doi = "10.1109/ASPDAC.2018.8297290",
language = "English",
series = "Proceedings of the Asia and South Pacific Design Automation Conference, ASP-DAC",
publisher = "Institute of Electrical and Electronics Engineers Inc.",
pages = "105--110",
booktitle = "ASP-DAC 2018 - 23rd Asia and South Pacific Design Automation Conference, Proceedings",
address = "United States",
note = "null ; Conference date: 22-01-2018 Through 25-01-2018",
}