TY - GEN
T1 - Efficient Communication-Computation Tradeoff for Split Computing
T2 - 2023 IEEE Global Communications Conference, GLOBECOM 2023
AU - Cao, Yang
AU - Lien, Shao Yu
AU - Yeh, Cheng Hao
AU - Liang, Ying Chang
AU - Niyato, Dusit
N1 - Publisher Copyright:
© 2023 IEEE.
PY - 2023
Y1 - 2023
N2 - Splitting the computation loads of a neural network (NN) training task to multiple stations, split computing has been the most promising technology to sustain high-accuracy model for resource-constrained user equipments (UEs) to empower real-time intelligent services. Nevertheless, different communication link variations and computation capabilities in different stations (including UE and servers) render the overall performance optimization in split computing a critical challenge. In this case, different stations should be able to infer the others' communication/computation capabilities to distributively decide the optimum splitting points of an NN. To this end, in this paper, we propose a multi-tier deep reinforcement learning (DRL) scheme for split computing, by which the UE and edge server can collaboratively and adaptively determine their splitting points and computation resources to optimize the long-term overall training latency through tackling different time-scale sub-optimizations in a sequential manner. With the image recognition task as experimental example, comprehensive simulations are conducted to justify the performances in terms of training latency, model accuracy and energy consumption of the proposed scheme for split computing.
AB - Splitting the computation loads of a neural network (NN) training task to multiple stations, split computing has been the most promising technology to sustain high-accuracy model for resource-constrained user equipments (UEs) to empower real-time intelligent services. Nevertheless, different communication link variations and computation capabilities in different stations (including UE and servers) render the overall performance optimization in split computing a critical challenge. In this case, different stations should be able to infer the others' communication/computation capabilities to distributively decide the optimum splitting points of an NN. To this end, in this paper, we propose a multi-tier deep reinforcement learning (DRL) scheme for split computing, by which the UE and edge server can collaboratively and adaptively determine their splitting points and computation resources to optimize the long-term overall training latency through tackling different time-scale sub-optimizations in a sequential manner. With the image recognition task as experimental example, comprehensive simulations are conducted to justify the performances in terms of training latency, model accuracy and energy consumption of the proposed scheme for split computing.
UR - http://www.scopus.com/inward/record.url?scp=85187401398&partnerID=8YFLogxK
U2 - 10.1109/GLOBECOM54140.2023.10437522
DO - 10.1109/GLOBECOM54140.2023.10437522
M3 - Conference contribution
AN - SCOPUS:85187401398
T3 - Proceedings - IEEE Global Communications Conference, GLOBECOM
SP - 176
EP - 181
BT - GLOBECOM 2023 - 2023 IEEE Global Communications Conference
PB - Institute of Electrical and Electronics Engineers Inc.
Y2 - 4 December 2023 through 8 December 2023
ER -