2020
DOI: 10.3390/app10165574
|View full text |Cite
|
Sign up to set email alerts
|

Deep Reinforcement Learning with Interactive Feedback in a Human–Robot Environment

Abstract: Robots are extending their presence in domestic environments every day, it being more common to see them carrying out tasks in home scenarios. In the future, robots are expected to increasingly perform more complex tasks and, therefore, be able to acquire experience from different sources as quickly as possible. A plausible approach to address this issue is interactive feedback, where a trainer advises a learner on which actions should be taken from specific states to speed up the learning process. Moreover, d… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

0
25
0
4

Year Published

2020
2020
2022
2022

Publication Types

Select...
9

Relationship

3
6

Authors

Journals

citations
Cited by 37 publications
(29 citation statements)
references
References 28 publications
0
25
0
4
Order By: Relevance
“…Human-sourced information has shown great potential due to its breadth, depth, and availability [ 9 ]. ARL agents that interact particularly with humans during operation are known as interactive agents, these agents have shown large improvements over unassisted agents [ 10 , 11 , 12 , 13 ].…”
Section: Introductionmentioning
confidence: 99%
“…Human-sourced information has shown great potential due to its breadth, depth, and availability [ 9 ]. ARL agents that interact particularly with humans during operation are known as interactive agents, these agents have shown large improvements over unassisted agents [ 10 , 11 , 12 , 13 ].…”
Section: Introductionmentioning
confidence: 99%
“…We implement a simulated robot arm for the task of organizing different objects [57]. The objective is to classify six geometric figures with different shapes and colors.…”
Section: B Simulated Robotic Armmentioning
confidence: 99%
“…until x t is terminal 17: end for and a simulated robotic arm for the task of organizing different objects [57].…”
mentioning
confidence: 99%
“…DRL has the advantages of not requiring environmental maps, strong learning capabilities, and high dynamic adaptability. Its deep network can successfully process high-dimensional information from sensors, and its reinforcement learning mechanism can perform continuous decision-making tasks in complex environments [ 16 , 17 , 18 ]. The successfully trained network model can directly generate optimal control commands for the lunar rover based on sensor information, omitting the complex environment reconstruction and sensing steps of traditional algorithms, and is very suitable for dynamic planning tasks such as lunar exploration with the unknown environment and limited on-board computing resources.…”
Section: Introductionmentioning
confidence: 99%