2017
DOI: 10.1016/j.robot.2017.03.003
|View full text |Cite
|
Sign up to set email alerts
|

Improving the speed of convergence of multi-agent Q-learning for cooperative task-planning by a robot-team

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
14
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
4
2
2

Relationship

0
8

Authors

Journals

citations
Cited by 30 publications
(14 citation statements)
references
References 20 publications
0
14
0
Order By: Relevance
“…A comparative analysis of the proposed approach is done with other approaches mentioned in the literature. Ability and complexity (Sadhu and Konar 2017 ; Gadaleta 2017 ) both are analyzed with respect to the utilized time space and effectiveness. Time required to complete the execution process is referred as time complexity.…”
Section: Results Analysismentioning
confidence: 99%
“…A comparative analysis of the proposed approach is done with other approaches mentioned in the literature. Ability and complexity (Sadhu and Konar 2017 ; Gadaleta 2017 ) both are analyzed with respect to the utilized time space and effectiveness. Time required to complete the execution process is referred as time complexity.…”
Section: Results Analysismentioning
confidence: 99%
“…In most cases, Q-function with multistate is used to better learn the environment (Fernandez-Gauna et al, 2013;Sadhu and Konar, 2017;Chai and Hayashibe, 2020). Particularly, this paper (Chai and Hayashibe, 2020) has explored deep RL for motion generation in a simulated environment.…”
Section: Q-learning Structurementioning
confidence: 99%
“…Busoniu et al [2] provides a comprehensive survey of multi-agent reinforcement learning (MARL) techniques for fully cooperative, fully competitive, and mixed tasks which focuses on autonomous multiple agents learning to solve complex tasks online using learning strategies based on dynamic programming and temporal-difference RL. Sadhu et al [3] proposed an algorithm for MAQL with two appealing features which were unavailable in convention Q learning. To begin with, an agent only needs to adapt one Q-table in joint state-action space during the learning process, as opposed to m joint Q-tables for, m agents in CQL.…”
Section: Introductionmentioning
confidence: 99%
“…This helps in the convergence of the process. Sadhu et al [3] has efficient convergence as it implements team-goal exploration and joint action selection for the specified state. Chaplot et al [6] focuses on Active Neural SLAM, a modular navigational model.…”
Section: Introductionmentioning
confidence: 99%