检索规则说明:AND代表“并且”;OR代表“或者”;NOT代表“不包含”;(注意必须大写,运算符两边需空一格)
检 索 范 例 :范例一: (K=图书馆学 OR K=情报学) AND A=范并思 范例二:J=计算机应用与软件 AND (U=C++ OR U=Basic) NOT M=Visual
作 者:Yong-feng Li Jing-ping Shi Wei Jiang Wei-guo Zhang Yong-xi Lyu
机构地区:[1]School of Automation,Northwestern Polytechnical University,Xi'an 710129,China [2]Shaanxi Province Key Laboratory of Flight Control and Simulation Technology,Xi'an 710129,China
出 处:《Defence Technology(防务技术)》2022年第9期1697-1714,共18页Defence Technology
基 金:supported by the National Natural Science Foundation of China (No. 61573286);the Aeronautical Science Foundation of China (No. 20180753006);the Fundamental Research Funds for the Central Universities (3102019ZDHKY07);the Natural Science Foundation of Shaanxi Province (2019JM-163, 2020JQ-218);the Shaanxi Province Key Laboratory of Flight Control and Simulation Technology。
摘 要:To solve the problem of realizing autonomous aerial combat decision-making for unmanned combat aerial vehicles(UCAVs) rapidly and accurately in an uncertain environment, this paper proposes a decision-making method based on an improved deep reinforcement learning(DRL) algorithm: the multistep double deep Q-network(MS-DDQN) algorithm. First, a six-degree-of-freedom UCAV model based on an aircraft control system is established on a simulation platform, and the situation assessment functions of the UCAV and its target are established by considering their angles, altitudes, environments, missile attack performances, and UCAV performance. By controlling the flight path angle, roll angle, and flight velocity, 27 common basic actions are designed. On this basis, aiming to overcome the defects of traditional DRL in terms of training speed and convergence speed, the improved MS-DDQN method is introduced to incorporate the final return value into the previous steps. Finally, the pre-training learning model is used as the starting point for the second learning model to simulate the UCAV aerial combat decision-making process based on the basic training method, which helps to shorten the training time and improve the learning efficiency. The improved DRL algorithm significantly accelerates the training speed and estimates the target value more accurately during training, and it can be applied to aerial combat decision-making.
关 键 词:Unmanned combat aerial vehicle Aerial combat decision Multi-step double deep Q-network Six-degree-of-freedom Aerial combat maneuver library
分 类 号:V249.1[航空宇航科学与技术—飞行器设计] V271.4
正在载入数据...
正在载入数据...
正在载入数据...
正在载入数据...
正在载入数据...
正在载入数据...
正在载入数据...
正在链接到云南高校图书馆文献保障联盟下载...
云南高校图书馆联盟文献共享服务平台 版权所有©
您的IP:3.148.232.123