2021
DOI: 10.1109/access.2021.3060426
|View full text |Cite
|
Sign up to set email alerts
|

Application of Deep Reinforcement Learning in Maneuver Planning of Beyond-Visual-Range Air Combat

Abstract: Beyond-visual-range (BVR) engagement becomes more and more popular in the modern air battlefield. The key and difficulty for pilots in the fight is maneuver planning, which reflects the tactical decision-making capacity of the both sides and determinates success or failure. In this paper, we propose an intelligent maneuver planning method for BVR combat with using an improved deep Q network (DQN). First, a basic combat environment builds, which mainly includes flight motion model, relative motion model and mis… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
5

Citation Types

0
27
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
7
2
1

Relationship

0
10

Authors

Journals

citations
Cited by 51 publications
(27 citation statements)
references
References 19 publications
0
27
0
Order By: Relevance
“…A comprehensive mathematical model of air combat confrontation is established to effectively improves the fidelity of confrontation state. Through a large number of machine-to-machine confrontation simulation experiments with various initial states based on the MATLAB platform [22,23], an analysis of the maneuvering decision process is performed; the results verify that the model can autonomously output reasonable maneuvers for a certain goal and ensure highquality results for fighter missile attack decisions and kill efficiency calculations. The proposed dynamic method based on influence diagrams can be used to realistically assess a situation involving BVR air combat, and the superiority and killing benefit calculations are more accurate than those based on static methods; therefore, this method can accurately relate equipment contributions and the roles of equipment in a system.…”
Section: Introductionmentioning
confidence: 96%
“…A comprehensive mathematical model of air combat confrontation is established to effectively improves the fidelity of confrontation state. Through a large number of machine-to-machine confrontation simulation experiments with various initial states based on the MATLAB platform [22,23], an analysis of the maneuvering decision process is performed; the results verify that the model can autonomously output reasonable maneuvers for a certain goal and ensure highquality results for fighter missile attack decisions and kill efficiency calculations. The proposed dynamic method based on influence diagrams can be used to realistically assess a situation involving BVR air combat, and the superiority and killing benefit calculations are more accurate than those based on static methods; therefore, this method can accurately relate equipment contributions and the roles of equipment in a system.…”
Section: Introductionmentioning
confidence: 96%
“…However, when the space dimension increases, it faces the explosion problem of the state space and action space, and its decision accuracy is bound to be affected by fuzziness. Many scholars have adopted different DQN algorithms such as DQN [9,32], LSTM-DQN [33], and MS-DDQN [34] to realize the maneuver learning of UAV in short-range air combat and solved the decision problem of continuous state space. However, its action space basically adopts the form of maneuver action library, and the limited maneuver action library is difficult to reflect the maneuver action in actual air combat.…”
Section: Introductionmentioning
confidence: 99%
“…However, the maneuver library of this method only contains five maneuvers, which cannot meet the needs of air combat. Hu et al ( 2021 ) proposed to use the improved deep Q network (Mnih et al, 2015 ) for maneuver decisions in autonomous air combat, constructed the relative motion model, missile attack model, maneuver decision-making framework, designed the reward function for training agents, and replaced the strategy network in deep Q network with the perception situation layer and value fitting layer. This method improves the winning rate of air combat, but the maneuver library is relatively simple and difficult to meet the needs of air combat.…”
Section: Introductionmentioning
confidence: 99%