2017
DOI: 10.1007/s10846-017-0696-1
|View full text |Cite
|
Sign up to set email alerts
|

Learning to Perform a Perched Landing on the Ground Using Deep Reinforcement Learning

Abstract: A UAV with a variable sweep wing has the potential to perform a perched landing on the ground by achieving high pitch rates to take advantage of dynamic stall. This study focuses on the generation and evaluation of a trajectory to perform a perched landing on the ground using a non-linear constraint optimiser (Interior Point OPTimizer) and a Deep Q-Network (DQN). The trajectory is generated using a numerical model that characterises the dynamics of a UAV with a variable sweep wing which was developed through w… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4

Citation Types

0
34
0

Year Published

2018
2018
2024
2024

Publication Types

Select...
8
1

Relationship

0
9

Authors

Journals

citations
Cited by 35 publications
(34 citation statements)
references
References 7 publications
0
34
0
Order By: Relevance
“…Some RL methods have been proposed to solve aircraft guidance problems. [18][19][20][21][22] An RL agent is developed for guiding a powered UAV from one thermal location to another quickly and efficiently by controlling bank angle 18 in the x-y plane. To solve guidance problem with many conflicting objectives, a voting Q-learning algorithm is proposed.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…Some RL methods have been proposed to solve aircraft guidance problems. [18][19][20][21][22] An RL agent is developed for guiding a powered UAV from one thermal location to another quickly and efficiently by controlling bank angle 18 in the x-y plane. To solve guidance problem with many conflicting objectives, a voting Q-learning algorithm is proposed.…”
Section: Introductionmentioning
confidence: 99%
“…A DQN algorithm is used to generate a trajectory to perform a perched landing on the ground. 22 In this DQN algorithm, noise is added to the numerical model of airspeed in the training environment, which is more in line with the actual scenario.…”
Section: Introductionmentioning
confidence: 99%
“…A similar approach was proposed in another study [ 13 ], which used DRL to train the aircraft by heading commands and constant speed to guide the aircraft. A trajectory generating method was proposed by using a DQN algorithm to perform a perched landing on the ground [ 14 ]. In the above DQN algorithm, noise is considered by the model, which is more in line with the actual scenario in the training process.…”
Section: Introductionmentioning
confidence: 99%
“…DRL has been utilized in many decision-making fields, such as video games [26,27], board games [28,29], and robot control [30] and obtained great achievements of human level or superhuman performance. For aircraft guidance research, Waldock et al [31] proposed a DQN method to generate a trajectory to perform perched landing on the ground. Alejandro et al [32] proposed a DRL strategy for autonomous landing of the UAV on a moving platform.…”
Section: Introductionmentioning
confidence: 99%