2021
DOI: 10.1109/access.2021.3118109
|View full text |Cite
|
Sign up to set email alerts
|

Automatic Curriculum Design for Object Transportation Based on Deep Reinforcement Learning

Abstract: This paper presents an automatic curriculum learning (ACL) method for object transportation based on deep reinforcement learning (DRL). Previous studies on object transportation using DRL have a sparse reward problem that an agent receives a rare reward for only the transportation completion of an object. Generally, curriculum learning (CL) has been used to solve the sparse reward problem. However, the conventional CL methods should be manually designed by users, which is difficult and tedious work. Moreover, … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
6
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
5

Relationship

2
3

Authors

Journals

citations
Cited by 5 publications
(6 citation statements)
references
References 54 publications
0
6
0
Order By: Relevance
“…The other limitation is how much we modify the training environment that is built based on the validation environment, i.e., when generating a training environment, how much variance in the Gaussian distribution related to the number of obstacles, positions, and radii should be adjusted? Recent research has proposed to automatically adjust the difficulty of curriculum environments based on the robot's learning performance in the training environment [39]. Integrating such advances into our study by combining methods for generating appropriate subtasks and automatically generating curriculum environments for each task could address the limitations of our approach.…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…The other limitation is how much we modify the training environment that is built based on the validation environment, i.e., when generating a training environment, how much variance in the Gaussian distribution related to the number of obstacles, positions, and radii should be adjusted? Recent research has proposed to automatically adjust the difficulty of curriculum environments based on the robot's learning performance in the training environment [39]. Integrating such advances into our study by combining methods for generating appropriate subtasks and automatically generating curriculum environments for each task could address the limitations of our approach.…”
Section: Discussionmentioning
confidence: 99%
“…The robot can dynamically adjust the task difficulty based on the robot's performance, which provides a more appropriate curriculum design compared to the traditional methods [38]. Eoh and Park [39] proposed an automatic curriculum method for object transportation by generating the difficulty map. The curriculum is generated adaptively from easy to difficult environments for object transportation.…”
Section: B Curriculum Learningmentioning
confidence: 99%
“…Zhang et al propose a deep reinforcement learning algorithm in which each robot use a deep Q-learning controller to transport the oversized object for a single robot [92]. Eoh et al find that a deep reinforcement learning algorithm takes a long time to train a policy, and random action is challenging to train a satisfactory policy [96]. They use curriculum-based learning methods and propose region-growing and single-to multi-robot curricula that raise the success rate of object transportation tasks.…”
Section: ) Reinforcement Learning Based Controlmentioning
confidence: 99%
“…Each decentralized Q-net is trained with the help of a centralized Q-net. Eoh and Park [ 29 ] proposed a curriculum-learning-based object transportation method using difficulty map generation and an adaptive determination of the episode size. Shibata et al [ 30 ] presented a DRL-based multi-robot transportation method using an event-triggered communication and consensus-based control.…”
Section: Related Workmentioning
confidence: 99%