Journal of Systems Engineering and Electronics ›› 2026, Vol. 37 ›› Issue (1): 211-224.doi: 10.23919/JSEE.2026.000016

• SYSTEMS ENGINEERING • Previous Articles     Next Articles

A formation pursuit method integrated coordinated reciprocity for enhanced capture

Xiaoyu XING1(), Haoxiang XIA1,2,3,*()   

  1. 1School of Economics and Management, Dalian University of Technology, Dalian 116024, China
    2Key Laboratory of Social Computing and Cognitive Intelligence, Ministry of Education of China, Dalian 116024, China
    3Institute of Advanced Intelligence, Dalian University of Technology, Dalian 116024, China
  • Received:2024-03-27 Accepted:2026-01-05 Online:2026-02-18 Published:2026-03-09
  • Contact: Haoxiang XIA E-mail:xiaoyuxing1109@163.com;hxxia@dlut.edu.cn
  • About author:
    XING Xiaoyu was born in 1994. He received his M.S. degree in management science and engineering from Dalian University of Technology in 2020. He is currently pursuing his Ph.D. degree in management science and engineering at Dalian University of Technology. His research interests include machine learning, deep reinforcement learning, and multi-agent cooperative control. E-mail: xiaoyuxing1109@163.com

    XIA Haoxiang was born in 1972. He received his Ph.D. degree in systems engineering from the Institute of Systems Engineering, Dalian University of Technology, Dalian, China in 1998. He is a professor at the School of Economics and Management, Dalian University of Technology, Dalian, China. His current research interests are complex systems and evolutionary games. E-mail: hxxia@dlut.edu.cn
  • Supported by:
    This work was supported by the National Natural Science Foundation of China (72371052; 71871042).

Abstract:

Cooperative pursuit poses challenges across natural, social, and technical systems, particularly when decentralized, slow-speed pursuers attempt to capture a high-speed evader with limited observation. Most existing contributions place the focus on the greedy pursuit of the evader, overlooking potential collaborations among pursuers. To tackle this issue, a decision-making framework of multi-agent coordinated reciprocity formation pursuit (MACRFP) via deep reinforcement learning is introduced. This framework integrates the actor-critic algorithm with the coordinated reciprocity mechanism to enhance the capability of capturing a faster evader. Initially, a local perception model is created by utilizing a cellular network to simulate limitations caused by obstacles. Next, the formation coalition of pursuit is guided by the Cartesian Oval, enabling dispersed pursuers to create a siege against the faster evader. Furthermore, a coordinated reciprocity model based on the coordination graph and the attention-based graph neural networks is developed, addressing the global coordination problem by estimating a reciprocity coefficient to adjust agents’ rewards. Numerical simulations demonstrate the emergence of cooperative behaviors in cooperative besiegement, target tracking, and intelligent interception during the pursuit, indicating that the proposed algorithm enhances the feasibility and effectiveness of capturing a fast-escaping target by integrating coordinated reciprocity and coalition formation.

Key words: multi-agent system, reinforcement learning, cooperative pursuit, coordinated reciprocity