欢迎访问《兵工学报》官方网站,今天是 分享到:

兵工学报 ›› 2022, Vol. 43 ›› Issue (S1): 74-81.doi: 10.12382/bgxb.2022.A005

• 论文 • 上一篇    下一篇

国产化环境下基于强化学习的地空协同作战仿真

李理, 李旭光, 郭凯杰, 史超, 陈昭文   

  1. (中国兵器工业计算机应用技术研究所 车辆综合电子研发部, 北京 100089)
  • 上线日期:2022-06-28
  • 作者简介:李理(1993—),女,工程师,硕士。E-mail:qqgirllily@163.com
  • 基金资助:
    兵器联合基金项目(6141B012301)

Simulation of Ground-air Cooperative Combat Based on Reinforcement Learning in Localization Environment

LI Li, LI Xuguang, GUO Kaijie, SHI Chao, CHEN Zhaowen   

  1. (Departament of Vehicle Integrated Electronics Research and Development,Institute of Computer Application Technology, Norinco Group, Beijing 100089, China)
  • Online:2022-06-28

摘要: 以未来战场无人地空协同作战为需求牵引,面对军事领域实战场景匮乏、训练数据不足的实际问题,聚焦仿真环境下的深度强化学习方法,实现地空协同作战仿真中多智能体决策模型。在飞腾CPU和昆仑K200硬件平台与麒麟V10操作系统环境下搭建虚拟仿真环境,设置仿真环境状态表征、各智能体动作空间及奖励机制,构建基于深度确定性策略梯度算法的多智能体模型(MADDPG),通过仿真实验验证采用MADDPG算法能够使奖励值在地空协同作战仿真场景中逐渐收敛,从而证明该模型应用于地空协同作战的决策有效性。

关键词: 地空协同作战, 强化学习, 深度确定性策略梯度算法, 多智能体模型, 国产化环境

Abstract: For the actual problems of lack of actual combat scenes and insufficient training data in the military field, the deep reinforcement learning method is used to realize a multi-agent decision-making model for the unmanned ground and air cooperative combat simulation. A virtual simulation environment is built using Phytium CPU and Kunlun K200 NPU as hardware platform and Kylin V10 operating system as software environment. The simulation environment state representation,agents’ action space and rewards mechanism are set, and a multi-agent decision-making model based on the deep deterministic policy gradient (MADDPG) algorithm is established. Simulation experiments verified that MADDPG algorithm can make reward value converge gradually in ground-air cooperative combat simulation scenarios, thus proving the effectiveness of MADDPG algorithm in the simulation scene of the ground-air cooperative combat.

Key words: ground-aircooperativecombat, reinforcementlearning, deepdeterministicpolicygradientalgorithm, multi-agentmodel, localizationenvironment

中图分类号: