2019 International Conference of Artificial Intelligence and Information Technology (ICAIIT) 2019
DOI: 10.1109/icaiit.2019.8834601
|View full text |Cite
|
Sign up to set email alerts
|

Reinforcement Point and Fuzzy Input Design of Fuzzy Q-Learning for Mobile Robot Navigation System

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
10
0
1

Year Published

2020
2020
2023
2023

Publication Types

Select...
3
3
3

Relationship

0
9

Authors

Journals

citations
Cited by 16 publications
(12 citation statements)
references
References 4 publications
1
10
0
1
Order By: Relevance
“…From the simulation results and experimental results, compared with some other algorithms: Q-learning, DDPG algorithm is better than DQN, Q-learning in terms of value accuracy and control strategy are presented in [13,16], this is also consistent with the DDPG algorithm that the authors have proposed in the paper. Accelerated learning technology and rapid action processing in large environments, can be used to achieve action status maps and meet the mobility needs of mobile robots.…”
Section: Fig 6 Results Of Navigating Robot Turtlebot Done On Rvizsupporting
confidence: 73%
See 1 more Smart Citation
“…From the simulation results and experimental results, compared with some other algorithms: Q-learning, DDPG algorithm is better than DQN, Q-learning in terms of value accuracy and control strategy are presented in [13,16], this is also consistent with the DDPG algorithm that the authors have proposed in the paper. Accelerated learning technology and rapid action processing in large environments, can be used to achieve action status maps and meet the mobility needs of mobile robots.…”
Section: Fig 6 Results Of Navigating Robot Turtlebot Done On Rvizsupporting
confidence: 73%
“…Therefore, to build a complete DDPG algorithm, it is always necessary to meet the needs of selecting a control action to a robot, executing the action, receiving rewards, storing and sampling to train the algorithm, calculating of the target function. Subsequently updating the model parameters by minimizing the loss function on all selected samples, followed by selecting the method to update the target neural network parameters, and finally updating the environmental discovery coefficient during the control process [3,5,6,15,16].…”
Section: B the Robot Navigation Using Ddpg Algorithmmentioning
confidence: 99%
“…Dalam perencanaan jalur robot dengan hambatan statis, perlu perhitungan yang pasti untuk menghindari hambatan sesuai dengan jalur yang telah direncanakan [4]. Untuk mobil robot yang memiliki heading, penelitian ini fokus pada seberapa optimal robot menghindari hambatan dinamis supaya jalur yang ditempuh adalah minimal [5]. Artikel ini disusun berdasarkan urutan sebagai berikut.…”
Section: Pendahuluanunclassified
“…There are some Fuzzy adaptations on [24] work like [25] where the Q-functions and action selection strategy are inferred from Fuzzy rules. Also, in order to reduce the number of states needed to shape an MDP model for mobile robots that avoid obstacles, [26] suggests a Fuzzy technique. Because the mobile robot may encounter an infinite number of different conditions.…”
Section: Introductionmentioning
confidence: 99%