Journal of Systems Engineering and Electronics ›› 2026, Vol. 37 ›› Issue (1): 242-256.doi: 10.23919/JSEE.2023.000165

• SYSTEMS ENGINEERING • Previous Articles     Next Articles

Optimal competitive resource assignment in two-stage Colonel Blotto game with Lanchester-type attrition

Weilin YUAN(), Shaofei CHEN(), Zhenzhen HU(), Xiang JI(), Lina LU(), Xiaolong SU(), Jing CHEN()   

  • Received:2022-12-12 Online:2026-02-18 Published:2026-03-11
  • Contact: Shaofei CHEN E-mail:yuanweilin12@nudt.edu.cn;chensf005@163.com;hzzmail@163.com;jixiang14@nudt.edu.cn;lulina16@nudt.edu.cn;xiaolongsu@nudt.edu.cn;Chenjing001@vip.sina.com
  • About author:
    YUAN Weilin was born in 1994. He received his B.S, M.S, and Ph.D. degrees from control science and engineering from the National University of Defense Technology, Changsha, China, in 2012, 2016 and 2023, respectively. He is a lecturer in National University of Defense Technology. His research interests include cognitive decision-making, intelligent gaming, opponent modeling, reinforcement learning, and multi-agent system. E-mail: yuanweilin12@nudt.edu.cn

    CHEN Shaofei was born in 1987. He received his B.S. degree from Harbin Institute of Technology, Harbin, China, in 2009, M.S. and Ph.D. degrees in control science and engineering from National University of Defense Technology, Changsha, in 2011 and 2016, respectively. Since 2019, he has been an associate professor with the College of Intelligence Science and Technology, National University of Defense Technology. His research interests include artificial intelligence, multiagent system, and reinforcement learning. E-mail: chensf005@163.com

    HU Zhenzhen born in 1984. He received his B.S. degree in thermal energy and power engineering from Shanghai Jiaotong University in 2006 and M.S. degree in fluid mechanics at China Aerodynamics Research and Development Center in 2009. He is currently pursuing his Ph.D. degree in control science and engineering at the National University of Defense Technology, Changsha, China. His current research interests include artificial intelligence, opponent modeling, and game theory. E-mail: hzzmail@163.com

    JI Xiang was born in 1991. He received his M.S, and Ph.D. degrees in control science and engineering from the National University of Defense Technology, Changsha, China, in 2014 and 2022, respectively. He is a lecturer in National University of Defense Technology. His research interests include swarm intelligence, complex network, and intelligent gaming. E-mail: jixiang14@nudt.edu.cn

    LU Lina born in 1984. She received her Ph.D. degree in control science and engineering from the National University of Defense Technology, Changsha, China, in 2020. Since 2020, she has been a lecturer with the College of Intelligence Science and Technology, National University of Defense Technology. Her current research interests include reinforcement learning, opponent modeling, and complex network. E-mail: lulina16@nudt.edu.cn

    SU Xiaolong born in 2000. He received his B.S. degree from the National University of Defense Technology, Changsha, China, in 2021. He is currently pursuing his M.S. degree in control science and engineering at the National University of Defense Technology, Changsha, China. His research interests include game theory, intelligent gaming, and reinforcement learning. E-mail: xiaolongsu@nudt.edu.cn

    CHEN Jing was born in 1972. He received his M.S. and Ph.D. degrees in control science and engineering from the National University of Defense Technology, Changsha, China, in 1993 and 1999, respectively. He is a full professor with the Department of Intelligence Science and Technology, College of Intelligence Science and Technology, National University of Defense Technology. His current research interests include artificial intelligence, intelligence control, and unmanned vehicle mission planning. E-mail: Chenjing001@vip.sina.com
  • Supported by:
    This work was supported by the National Natural Science Foundation of China (61702528; 61806212; 62173336).

Abstract:

In strategic decision-making tasks, determining how to assign limited costly resource towards the defender and the attacker is a central problem. However, it is hard for pre-allocated resource assignment to adapt to dynamic fighting scenarios, and exists situations where the scenario and rule of the Colonel Blotto (CB) game are too restrictive in real world. To address these issues, a support stage is added as supplementary for pre-allocated results, in which a novel two-stage competitive resource assignment problem is formulated based on CB game and stochastic Lanchester equation (SLE). Further, the force attrition in these two stages is formulated as a stochastic progress to consider the complex fighting progress, including the case that the player with fewer resources defeats the player with more resources and wins the battlefield. For solving this two-stage resource assignment problem, nested solving and no-regret learning are proposed to search the optimal resource assignment strategies. Numerical experiments are taken to analyze the effectiveness of the proposed model and study the assignment strategies in various cases.

Key words: resource assignment, Colonel Blotto (CB) game, stochastic Lanchester equation (SLE), regret match