TY - GEN
T1 - A Reconfigurable Deep Neural Network on Chip Design with Flexible Convolutional Operations
AU - Chen, Kun Chih
AU - Liao, Yi Sheng
N1 - Publisher Copyright:
© 2022 IEEE.
PY - 2022
Y1 - 2022
N2 - The designs of deep neural network (DNN) accelerators have gradually gained attention due to the increased demand for real-Time AI applications. On the other hand, due to the diverse applications, kernel sizes and shapes for the involved convolutional operation in the target DNN model are not fixed. Therefore, it is necessary to design a reconfigurable DNN accelerator to cover different kernel sizes for convolutional operation in DNNs. However, due to the worst-case design policy, the designers usually select the largest kernel size as the design parameter to implement the DNN accelerator, which leads to lower hardware utilization. The reason is that the conventional array-based DNN design method restricts the efficiency of data delivery. Besides, the complicated data flow between neuron layers of DNN models counteracts the benefit of the involved data reuse method. To mitigate the design problems of complicated data flow on DNN accelerators, Network-on-Chip (NoC) interconnection has become an emerging technology to realize the Deep Neural Network on Chip (DNNoC). Compared with the conventional array-based DNN acceleration design, the DNNoC design supports flexible data flow, which leverages reconfigurable DNN accelerator implementations. In this work, we leverage the flexible NoC interconnection and propose a hybrid input/weight reuse method to reduce memory access. In addition, our proposed hybrid input/weight reuse method supports arbitrary kernel sizes for flexible convolutional operations. Compared with the related works, the proposed reconfigurable DNNoC with flexible convolutional operations helps to improve the utilization of computational capability in a PE by 1 % to 34 %, reduce memory access by 66% to 85%, which helps to improve 40% to 117% throughput.
AB - The designs of deep neural network (DNN) accelerators have gradually gained attention due to the increased demand for real-Time AI applications. On the other hand, due to the diverse applications, kernel sizes and shapes for the involved convolutional operation in the target DNN model are not fixed. Therefore, it is necessary to design a reconfigurable DNN accelerator to cover different kernel sizes for convolutional operation in DNNs. However, due to the worst-case design policy, the designers usually select the largest kernel size as the design parameter to implement the DNN accelerator, which leads to lower hardware utilization. The reason is that the conventional array-based DNN design method restricts the efficiency of data delivery. Besides, the complicated data flow between neuron layers of DNN models counteracts the benefit of the involved data reuse method. To mitigate the design problems of complicated data flow on DNN accelerators, Network-on-Chip (NoC) interconnection has become an emerging technology to realize the Deep Neural Network on Chip (DNNoC). Compared with the conventional array-based DNN acceleration design, the DNNoC design supports flexible data flow, which leverages reconfigurable DNN accelerator implementations. In this work, we leverage the flexible NoC interconnection and propose a hybrid input/weight reuse method to reduce memory access. In addition, our proposed hybrid input/weight reuse method supports arbitrary kernel sizes for flexible convolutional operations. Compared with the related works, the proposed reconfigurable DNNoC with flexible convolutional operations helps to improve the utilization of computational capability in a PE by 1 % to 34 %, reduce memory access by 66% to 85%, which helps to improve 40% to 117% throughput.
KW - DNN accelerator
KW - Network-on-Chip
KW - NoC
KW - deep neural network
UR - http://www.scopus.com/inward/record.url?scp=85141828974&partnerID=8YFLogxK
U2 - 10.1109/NoCArc57472.2022.9911283
DO - 10.1109/NoCArc57472.2022.9911283
M3 - Conference contribution
AN - SCOPUS:85141828974
T3 - 2022 15th International Workshop on Network on Chip Architectures, NoCArc 2022 - In conjunction with the 55th Annual IEEE/ACM International Symposium on Microarchitecture, MICRO 2022
BT - 2022 15th International Workshop on Network on Chip Architectures, NoCArc 2022 - In conjunction with the 55th Annual IEEE/ACM International Symposium on Microarchitecture, MICRO 2022
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 15th IEEE/ACM International Workshop on Network on Chip Architectures, NoCArc 2022
Y2 - 2 October 2022
ER -