TY - JOUR
T1 - Optimum splitting computing for DNN training through next generation smart networks
T2 - a multi-tier deep reinforcement learning approach
AU - Lien, Shao Yu
AU - Yeh, Cheng Hao
AU - Deng, Der Jiunn
N1 - Publisher Copyright:
© The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature 2024.
PY - 2024/4
Y1 - 2024/4
N2 - Deep neural networks (DNNs) involving massive neural nodes grouped into different neural layers have been a promising innovation for function approximation and inference, which have been widely applied to various vertical applications such as image recognition. However, the computing burdens to train a DNN model with a limited latency may not be affordable for the user equipment (UE), which consequently motivates the concept of splitting the computations of DNN layers to not only the edge server but also the cloud platform. Despite the availability of more computing resources, computing tasks with such split computing also suffer packet transmission unreliability, latency, and significant energy consumption. A practical scheme to optimally split the computations of DNN layers to the UE, edge, and cloud is thus urgently desired. To solve this optimization, we propose a multi-tier deep reinforcement learning (DRL) scheme for the UE and edge to distributively determine the splitting points to minimize the overall training latency while meeting the constraints of overall energy consumption and image recognition accuracy. The performance evaluation results show the outstanding performance of the proposed design as compared with state-of-the-art schemes, to fully justify the practicability in the next-generation smart networks.
AB - Deep neural networks (DNNs) involving massive neural nodes grouped into different neural layers have been a promising innovation for function approximation and inference, which have been widely applied to various vertical applications such as image recognition. However, the computing burdens to train a DNN model with a limited latency may not be affordable for the user equipment (UE), which consequently motivates the concept of splitting the computations of DNN layers to not only the edge server but also the cloud platform. Despite the availability of more computing resources, computing tasks with such split computing also suffer packet transmission unreliability, latency, and significant energy consumption. A practical scheme to optimally split the computations of DNN layers to the UE, edge, and cloud is thus urgently desired. To solve this optimization, we propose a multi-tier deep reinforcement learning (DRL) scheme for the UE and edge to distributively determine the splitting points to minimize the overall training latency while meeting the constraints of overall energy consumption and image recognition accuracy. The performance evaluation results show the outstanding performance of the proposed design as compared with state-of-the-art schemes, to fully justify the practicability in the next-generation smart networks.
KW - Deep neural networks (DNNs)
KW - Deep reinforcement learning (DRL)
KW - Sixth generation (6 G) Networks
KW - Split computing
UR - http://www.scopus.com/inward/record.url?scp=85181466036&partnerID=8YFLogxK
U2 - 10.1007/s11276-023-03600-5
DO - 10.1007/s11276-023-03600-5
M3 - Article
AN - SCOPUS:85181466036
SN - 1022-0038
VL - 30
SP - 1737
EP - 1751
JO - Wireless Networks
JF - Wireless Networks
IS - 3
ER -