2019
DOI: 10.1177/0036850419879024
|View full text |Cite
|
Sign up to set email alerts
|

Heuristic Q-learning based on experience replay for three-dimensional path planning of the unmanned aerial vehicle

Abstract: In order to solve the problem that the existing reinforcement learning algorithm is difficult to converge due to the excessive state space of the three-dimensional path planning of the unmanned aerial vehicle, this article proposes a reinforcement learning algorithm based on the heuristic function and the maximum average reward value of the experience replay mechanism. The knowledge of track performance is introduced to construct heuristic function to guide the unmanned aerial vehicles’ action selection and re… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
8
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
8
1

Relationship

0
9

Authors

Journals

citations
Cited by 17 publications
(8 citation statements)
references
References 11 publications
0
8
0
Order By: Relevance
“…RL algorithms for path planning are the most abundant in the state of the art. For example, Xie et al use the Q-Learning strategy for three-dimensional path planning [24]. The notion of Heuristic Q-Learning was introduced.…”
Section: Evolutionarymentioning
confidence: 99%
“…RL algorithms for path planning are the most abundant in the state of the art. For example, Xie et al use the Q-Learning strategy for three-dimensional path planning [24]. The notion of Heuristic Q-Learning was introduced.…”
Section: Evolutionarymentioning
confidence: 99%
“…Therefore, it is required to have strong adaptive ability to such uncertainty. RL provides a better idea for this kind of problem by using historical data to obtain the nonlinear function relationship between approximate fitting state and overall performance [21][22][23][24].…”
Section: Annotation Demo Sectionmentioning
confidence: 99%
“…They classified these approaches into five main categories. These categories include classical methods [32,33,34], heuristics [35,36,37,38,39,40], meta-heuristics [41,42,43], machine learning [44,45,46], and hybrid algorithms [47,48,49].…”
Section: Introductionmentioning
confidence: 99%