2015
DOI: 10.1016/j.neucom.2014.09.092
|View full text |Cite
|
Sign up to set email alerts
|

A hierarchical path planning approach based on A ⁎ and least-squares policy iteration for mobile robots

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
45
0

Year Published

2017
2017
2023
2023

Publication Types

Select...
6
2

Relationship

0
8

Authors

Journals

citations
Cited by 73 publications
(46 citation statements)
references
References 42 publications
1
45
0
Order By: Relevance
“…Given that it is difficult for an individual DRL method to solve the navigation problem, hierarchical approaches are widely researched in the literature [24][25][26]. Lei et al [26] combined A* and least-squares policy iteration for mobile robot navigation in complex environments. Aleksandra et al [25] integrated sampling-based path planning with reinforcement learning (RL) agents for indoor navigation and aerial cargo delivery.…”
Section: Deep Reinforcement Learningmentioning
confidence: 99%
See 2 more Smart Citations
“…Given that it is difficult for an individual DRL method to solve the navigation problem, hierarchical approaches are widely researched in the literature [24][25][26]. Lei et al [26] combined A* and least-squares policy iteration for mobile robot navigation in complex environments. Aleksandra et al [25] integrated sampling-based path planning with reinforcement learning (RL) agents for indoor navigation and aerial cargo delivery.…”
Section: Deep Reinforcement Learningmentioning
confidence: 99%
“…To sum up, JPS+ (P)-which we borrow from [32]-has obvious advantages over other high-level planners of hierarchical methods in the literature [24][25][26], in terms of its low precomputation costs and outstanding online performance. Moreover, it serves as an ideal for the location distribution of jump points, providing a DRL-based controller with meaningful subgoals that can completely throw the problem of local minima out of consideration.…”
Section: Global Path Planner Based On Jps+ (P)mentioning
confidence: 99%
See 1 more Smart Citation
“…In the first level, the algorithm employs grids to find a geometric path quickly, and several path points are selected as subgoals for the next level. In the second level, an approximate policy iteration algorithm denoted as leastsquares policy iteration (LSPI) is used to learn a near-optimal local planning policy that can generate smooth trajectories under the kinematic constraints of the robots [19]. Further improvements in the A * algorithm have been obtained by dividing the nodes generated by the algorithm into smaller steps, eliminating redundant nodes, and reducing the cost of step and handover [20].…”
Section: Introductionmentioning
confidence: 99%
“…11 In the RL framework, the robot can learn an optimal behavior policy through trialand-error by interacting with the working environment. Zuo et al 11 propose an improved A* algorithm which uses the least squares policy iteration to learn a near-optimal local planning policy that can generate smooth trajectories under kinematic constraints of the robot. Plaza et al 12 combine the RL strategies with cell-mapping techniques to solve the optimal-control problem for the car-like robot.…”
Section: Introductionmentioning
confidence: 99%