2017 36th Chinese Control Conference (CCC) 2017
DOI: 10.23919/chicc.2017.8027884
|View full text |Cite
|
Sign up to set email alerts
|

Q learning algorithm based UAV path learning and obstacle avoidence approach

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
35
0

Year Published

2018
2018
2024
2024

Publication Types

Select...
6
2
1

Relationship

0
9

Authors

Journals

citations
Cited by 82 publications
(35 citation statements)
references
References 16 publications
0
35
0
Order By: Relevance
“…To overcome the shortcomings of the methods described above, other studies [16][17][18][19] have proposed using reinforcement learning (RL), an online learning method, for UAV motion planning. Reinforcement learning can solve the UAV's AMP problem with navigating in an unknown environment, but as the environment becomes more complex and closer to reality, the "dimension curse" problem limits its development.…”
Section: Related Workmentioning
confidence: 99%
“…To overcome the shortcomings of the methods described above, other studies [16][17][18][19] have proposed using reinforcement learning (RL), an online learning method, for UAV motion planning. Reinforcement learning can solve the UAV's AMP problem with navigating in an unknown environment, but as the environment becomes more complex and closer to reality, the "dimension curse" problem limits its development.…”
Section: Related Workmentioning
confidence: 99%
“…To evaluate the performance of the proposed QL-GD, the comparison with CQL in mobile robot path planning was conducted. The experiment was carried out in simulated environment with various obstacle patterns in four different maps in MATLAB software by referring [29]. Size of all maps were 20 x 20 unit grid.…”
Section: Experiments Setupmentioning
confidence: 99%
“…In this method, it combines greedy strategy with Boltzmann strategy to improve exploration. Yijing et al [13] proposes an adaptive random exploration Q-learning method which designs the learning, escape and action modules of UAV respectively. Yan et al [14] initializes Q matrix with environment and target, and adds the function of avoiding repetitive actions in the initial stage of training.…”
Section: Introductionmentioning
confidence: 99%