%A Yaozhong ZHANG, Yike LI, Zhuoran WU, Jialin XU %T Deep reinforcement learning for UAV swarm rendezvous behavior %0 Journal Article %D 2023 %J Journal of Systems Engineering and Electronics %R 10.23919/JSEE.2023.000056 %P 360-373 %V 34 %N 2 %U {https://www.jseepub.com/CN/abstract/article_9180.shtml} %8 2023-04-18 %X

The unmanned aerial vehicle (UAV) swarm technology is one of the research hotspots in recent years. With the continuous improvement of autonomous intelligence of UAV, the swarm technology of UAV will become one of the main trends of UAV development in the future. This paper studies the behavior decision-making process of UAV swarm rendezvous task based on the double deep Q network (DDQN) algorithm. We design a guided reward function to effectively solve the problem of algorithm convergence caused by the sparse return problem in deep reinforcement learning (DRL) for the long period task. We also propose the concept of temporary storage area, optimizing the memory playback unit of the traditional DDQN algorithm, improving the convergence speed of the algorithm, and speeding up the training process of the algorithm. Different from traditional task environment, this paper establishes a continuous state-space task environment model to improve the authentication process of UAV task environment. Based on the DDQN algorithm, the collaborative tasks of UAV swarm in different task scenarios are trained. The experimental results validate that the DDQN algorithm is efficient in terms of training UAV swarm to complete the given collaborative tasks while meeting the requirements of UAV swarm for centralization and autonomy, and improving the intelligence of UAV swarm collaborative task execution. The simulation results show that after training, the proposed UAV swarm can carry out the rendezvous task well, and the success rate of the mission reaches 90%.