Acta Armamentarii ›› 2024, Vol. 45 ›› Issue (11): 3856-3867.doi: 10.12382/bgxb.2023.1048
Previous Articles Next Articles
LI Jiashen, WANG Xiaofang*(), LIN Hai
Received:
2024-01-26
Online:
2024-01-26
Contact:
WANG Xiaofang
CLC Number:
LI Jiashen, WANG Xiaofang, LIN Hai. Intelligent Penetration Policy for Hypersonic Cruise Missiles Based on Virtual Targets[J]. Acta Armamentarii, 2024, 45(11): 3856-3867.
Add to citation manager EndNote|Ris|BibTeX
参数 | Actor | Critic |
---|---|---|
输入层 | 7、10 | 7、10 |
激活函数1 | ReLU | ReLU |
隐藏层1 | 512 | 512 |
激活函数2 | ReLU | ReLU |
隐藏层2 | 256 | 256 |
激活函数3 | ReLU | ReLU |
α、β优化层 | 256 | |
激活函数4和5 | Softplus | |
输出层 | 2 | 1 |
Table 1 Hyperparameters of network architecture
参数 | Actor | Critic |
---|---|---|
输入层 | 7、10 | 7、10 |
激活函数1 | ReLU | ReLU |
隐藏层1 | 512 | 512 |
激活函数2 | ReLU | ReLU |
隐藏层2 | 256 | 256 |
激活函数3 | ReLU | ReLU |
α、β优化层 | 256 | |
激活函数4和5 | Softplus | |
输出层 | 2 | 1 |
参数 | 数值 |
---|---|
BatchSize | 140 |
MiniBatchSize | 70 |
训练回合数 | 2000 |
奖励折扣因子γ | 0.9 |
GAE平滑因子λ | 0.9 |
Epoch | 4 |
策略熵系数ke | 0.005 |
重要性采样权重裁剪因子ε | 0.2 |
初始学习率 | 0.015 |
学习率衰减节点 | 20、40、60、80 |
学习率衰减因子 | 0.6 |
Table 2 Hyperparameters for network training
参数 | 数值 |
---|---|
BatchSize | 140 |
MiniBatchSize | 70 |
训练回合数 | 2000 |
奖励折扣因子γ | 0.9 |
GAE平滑因子λ | 0.9 |
Epoch | 4 |
策略熵系数ke | 0.005 |
重要性采样权重裁剪因子ε | 0.2 |
初始学习率 | 0.015 |
学习率衰减节点 | 20、40、60、80 |
学习率衰减因子 | 0.6 |
[1] |
雷虎民, 骆长鑫, 周池军, 等. 临近空间防御作战拦截弹制导与控制关键技术综述[J]. 航空兵器, 2021, 28(2):1-10.
|
|
|
[2] |
汪丰麟, 李沁远, 范博, 等. 高超声速武器防御体系的发展现状与演进趋势[J]. 指挥与控制学报, 2022, 8(4):378-388.
|
|
|
[3] |
张荣升, 陈万春. THAAD增程型拦截弹预测制导方法[J]. 北京航空航天大学学报, 2021, 47(4):863-874.
|
|
|
[4] |
石安华, 李海燕, 石卫波, 等. 临近空间高超声速巡航飞行器红外特征[J]. 兵工学报, 2022, 43(4):796-803.
|
doi: 10.12382/bgxb.2021.0105 |
|
[5] |
|
[6] |
|
[7] |
郭行, 符文星, 付斌, 等. 吸气式高超声速飞行器巡航段突防弹道规划[J]. 宇航学报, 2017, 38(3):287-295.
|
|
|
[8] |
王雨琪, 宁国栋, 王晓峰, 等. 基于微分对策的临近空间飞行器机动突防策略[J]. 航空学报, 2020, 41(增刊2):724276.
|
|
|
[9] |
|
[10] |
|
[11] |
王芳, 林涛, 张克. 基于控制变量参数化的主动反拦截突防最优控制计算方法[J]. 航空学报, 2015, 36(6):2037-2046.
doi: 10.7527/S1000-6893.2014.0359 |
doi: 10.7527/S1000-6893.2014.0359 |
|
[12] |
樊博璇, 陈桂明, 林洪涛. 弹道导弹中段反应式机动突防规避策略[J]. 兵工学报, 2022, 43(1):69-78.
|
doi: 10.3969/j.issn.1000-1093.2022.01.008 |
|
[13] |
|
[14] |
张晚晴, 余文斌, 李静琳, 等. 基于纵程解析解的飞行器智能横程机动再入协同制导[J]. 兵工学报, 2021, 42(7): 1400-1411.
|
doi: 10.3969/j.issn.1000-1093.2021.07.007 |
|
[15] |
|
[16] |
吴杰, 张成, 李淼, 等. 基于凸优化和LQR的火箭返回轨迹跟踪制导[J]. 北京航空航天大学学报, 2022, 48(11):2270-2280.
|
|
|
[17] |
|
[18] |
王琦, 杨毅远, 江季. Easy RL:强化学习教程[M]. 北京: 人民邮电出版社, 2022:37-98.
|
|
|
[19] |
|
[20] |
|
[21] |
|
[1] | SUN Hao, LI Haiqing, LIANG Yan, MA Chaoxiong, WU Han. Dynamic Penetration Decision of Loitering Munition Group Based on Knowledge-assisted Reinforcement Learning [J]. Acta Armamentarii, 2024, 45(9): 3161-3176. |
[2] | HU Yanyang, HE Fan, BAI Chengchao. Cooperative Obstacle Avoidance Decision Method for the Terminal Guidance Phase of Hypersonic Vehicles [J]. Acta Armamentarii, 2024, 45(9): 3147-3160. |
[3] | CHEN Wenjie, CUI Xiaohong, WANG Binrui. Safety Optimal Tracking Control Algorithm and Robot Arm Simulation [J]. Acta Armamentarii, 2024, 45(8): 2688-2697. |
[4] | WANG Xiaolong, CHEN Yang, HU Mian, LI Xudong. Robot Path Planning for Persistent Monitoring Based on Improved Deep Q Networks [J]. Acta Armamentarii, 2024, 45(6): 1813-1823. |
[5] | FU Yanfang, LEI Kailin, WEI Jianing, CAO Zijian, YANG Bo, WANG Wei, SUN Zelong, LI Qinjie. A Hierarchical Multi-Agent Collaborative Decision-making Method Based on the Actor-critic Framework [J]. Acta Armamentarii, 2024, 45(10): 3385-3396. |
[6] | CAO Zijian, SUN Zelong, YAN Guochuang, FU Yanfang, YANG Bo, LI Qinjie, LEI Kailin, GAO Linghang. Simulation of Reinforcement Learning-based UAV Swarm Adversarial Strategy Deduction [J]. Acta Armamentarii, 2023, 44(S2): 126-134. |
[7] | LI Song, MA Zhuangzhuang, ZHANG Yunlin, SHAO Jinliang. Multi-agent Coverage Path Planning Based on Security Reinforcement Learning [J]. Acta Armamentarii, 2023, 44(S2): 101-113. |
[8] | YANG Jiaxiu, LI Xinkai, ZHANG Hongli, WANG Hao. Robust Tracking of Quadrotor UAVs Based on Integral Reinforcement Learning [J]. Acta Armamentarii, 2023, 44(9): 2802-2813. |
[9] | ZHANG Jiandong, WANG Dinghan, YANG Qiming, SHI Guoqing, LU Yi, ZHANG Yaozhong. Multi-Dimensional Decision-Making for UAV Air Combat Based on Hierarchical Reinforcement Learning [J]. Acta Armamentarii, 2023, 44(6): 1547-1563. |
[10] | LI Chao, WANG Ruixing, HUANG Jianzhong, JIANG Feilong, WEI Xuemei, SUN Yanxin. Autonomous Decision-making and Intelligent Collaboration of UAV Swarms Based on Reinforcement Learning with Sparse Rewards [J]. Acta Armamentarii, 2023, 44(6): 1537-1546. |
[11] | ZHENG Zexin, LI Wei, ZOU Kun, LI Yanfu. Anti-jamming Waveform Design of Ground-based Air Surveillance Radar Based on Reinforcement Learning [J]. Acta Armamentarii, 2023, 44(5): 1422-1430. |
[12] | LI Jiajian, SHI Yanjun, YANG Yu, LI Bo, ZHAO Xijun. Multi-agent Reinforcement Learning-based Offloading Decision for UAV Cluster Combat Tasks [J]. Acta Armamentarii, 2023, 44(11): 3295-3309. |
[13] | JIANG Yan, DING Yuyan, ZHANG Xinglong, XU Xin. A Human-machine Collaborative Control Algorithm for Intelligent Vehicles Based on Model Prediction and Policy Learning [J]. Acta Armamentarii, 2023, 44(11): 3465-3477. |
[14] | ZHAO Wenfei, CHEN Jian, WANG Yan, TENG Kenan. Dynamic Firepower Allocation for Cooperative Air Defense of Strategic Locations on the Sea Based on Reinforcement Learning [J]. Acta Armamentarii, 2023, 44(11): 3516-3528. |
Viewed | ||||||
Full text |
|
|||||
Abstract |
|
|||||