TY - GEN
T1 - Deep reinforcement learning based strategy for quadrotor uav pursuer and evader problem
AU - Chen, Dawei
AU - Wei, Yifei
AU - Wang, Li
AU - Hong, Choong Seon
AU - Wang, Li Chun
AU - Han, Zhu
N1 - Publisher Copyright:
© 2020 IEEE.
Copyright:
Copyright 2020 Elsevier B.V., All rights reserved.
PY - 2020/6
Y1 - 2020/6
N2 - In recent years, there have occurred many incidents that unmanned aerial vehicles (UAVs) in the field of national security. While in some situations, UAVs may be deployed simultaneously by different parties with opposite purposes, easily resulting in direct competitions against each other. In this case, how to use UAVs to pursue UAVs has become a hot spot. In order to analyze the behavior of UAV, building a realistic mathematical dynamic model is necessary. In this paper, we propose a Takagi-Sugeno (T-S) fuzzy control system based UAV dynamic model, which is exactly the same UAV control method in practice. To address the competition conundrum between UAVs, we formulate this problem into a pursuer-evader problem and leverage the reinforcement learning based machine learning method to solve this. The proposed deep Q network is based on traditional Q learning but able to address some deficiencies. Basically, deep Q network has three vital improvements: using neural network to describe the Q function, the architecture of double networks, and the experience replay. The simulation results show the correctness of our analysis and effectiveness of our proposed method.
AB - In recent years, there have occurred many incidents that unmanned aerial vehicles (UAVs) in the field of national security. While in some situations, UAVs may be deployed simultaneously by different parties with opposite purposes, easily resulting in direct competitions against each other. In this case, how to use UAVs to pursue UAVs has become a hot spot. In order to analyze the behavior of UAV, building a realistic mathematical dynamic model is necessary. In this paper, we propose a Takagi-Sugeno (T-S) fuzzy control system based UAV dynamic model, which is exactly the same UAV control method in practice. To address the competition conundrum between UAVs, we formulate this problem into a pursuer-evader problem and leverage the reinforcement learning based machine learning method to solve this. The proposed deep Q network is based on traditional Q learning but able to address some deficiencies. Basically, deep Q network has three vital improvements: using neural network to describe the Q function, the architecture of double networks, and the experience replay. The simulation results show the correctness of our analysis and effectiveness of our proposed method.
UR - http://www.scopus.com/inward/record.url?scp=85090275994&partnerID=8YFLogxK
U2 - 10.1109/ICCWorkshops49005.2020.9145456
DO - 10.1109/ICCWorkshops49005.2020.9145456
M3 - Conference contribution
AN - SCOPUS:85090275994
T3 - 2020 IEEE International Conference on Communications Workshops, ICC Workshops 2020 - Proceedings
BT - 2020 IEEE International Conference on Communications Workshops, ICC Workshops 2020 - Proceedings
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 2020 IEEE International Conference on Communications Workshops, ICC Workshops 2020
Y2 - 7 June 2020 through 11 June 2020
ER -