2016
DOI: 10.3906/elk-1311-129
|View full text |Cite
|
Sign up to set email alerts
|

Reinforcement learning-based mobile robot navigation

Abstract: In recent decades, reinforcement learning (RL) has been widely used in different research fields ranging from psychology to computer science. The unfeasibility of sampling all possibilities for continuous-state problems and the absence of an explicit teacher make RL algorithms preferable for supervised learning in the machine learning area, as the optimal control problem has become a popular subject of research. In this study, a system is proposed to solve mobile robot navigation by opting for the most popular… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
9
0
3

Year Published

2017
2017
2024
2024

Publication Types

Select...
7
1
1

Relationship

0
9

Authors

Journals

citations
Cited by 26 publications
(12 citation statements)
references
References 21 publications
0
9
0
3
Order By: Relevance
“…In the mobile robot framework structure in Figure 6, the agent is controlled by the supervisor controller, and the robot controller is executed by the agent, so that the robot operates in the warehouse environment and the necessary information is observed and collected [25][26][27].…”
Section: System Structurementioning
confidence: 99%
“…In the mobile robot framework structure in Figure 6, the agent is controlled by the supervisor controller, and the robot controller is executed by the agent, so that the robot operates in the warehouse environment and the necessary information is observed and collected [25][26][27].…”
Section: System Structurementioning
confidence: 99%
“…Typically, an agent will evaluate a state, and will then undertake an action either in an exploitative or exploratory manner thereafter and finally will receive an instant reward, while transitioning to a new state. Q-learning has tremendous success in robotics, especially in mobile robot navigation and obstacle avoidance [60,61]. In [62] the Dyna AI architecture was proposed to integrate both learning, and experience, based on online planning, as well as reactive execution in a stochastic environment.…”
Section: Generic Approaches To Uncertaintymentioning
confidence: 99%
“…Both behaviors need to produce dynamically feasible trajectories that are robust to many kinds of noise and inaccuracies, rely only on observations coming from primitive sensors, and avoid unexpected obstacles. In this paper, we present a reliable method to implement these navigation behaviors by learning 1 Google AI, Mountain View, CA 94043, USA lewispro,faust,mfiser,centaur@google.com * Authors contributed equally. end to end polices that directly map sensors to controls, and we show that these policies transfer from simulation to physical robots and new environments while robustly avoiding obstacles.…”
Section: Introductionmentioning
confidence: 99%