2020
DOI: 10.48550/arxiv.2006.13615
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Explainable robotic systems: Understanding goal-driven actions in a reinforcement learning scenario

Francisco Cruz,
Richard Dazeley,
Peter Vamplew
et al.

Abstract: Robotic systems are more present in our society everyday. In human-robot interaction scenarios, it is crucial that end-users develop trust in their robotic team-partners, in order to collaboratively complete a task. To increase trust, users demand more understanding about the decisions by the robot in particular situations. Recently, explainable robotic systems have emerged as an alternative focused not only on completing a task satisfactorily, but also in justifying, in a human-like manner, the reasons that l… Show more

Help me understand this report
View published versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
16
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
4
2

Relationship

3
3

Authors

Journals

citations
Cited by 7 publications
(16 citation statements)
references
References 38 publications
0
16
0
Order By: Relevance
“…To achieve this, each agent will compute the introspective confidence (ic a ) [38] of each action, which focuses on scaling the selected Q-value of an action towards the final goal using a logarithm transformation which computes the probability of success, in our case, of winning the game. The introspective confidence gives us a self-assessment of the agent's actions, based on its own game experience.…”
Section: Methodology and Experimental Protocolmentioning
confidence: 99%
“…To achieve this, each agent will compute the introspective confidence (ic a ) [38] of each action, which focuses on scaling the selected Q-value of an action towards the final goal using a logarithm transformation which computes the probability of success, in our case, of winning the game. The introspective confidence gives us a self-assessment of the agent's actions, based on its own game experience.…”
Section: Methodology and Experimental Protocolmentioning
confidence: 99%
“…• Partially observable environments: In practice, many RL problems are partially observable [148]. For instance, partial observabilities may occur in non-stationary environments [39] or in presence of stochastic transitions [149]. If the external information source does not have observations to clearly infer the current state in the environment may lead to giving incorrect assistance to the learner agent.…”
Section: Other Challengesmentioning
confidence: 99%
“…However, it is much more difficult to explain the agents' behavior in scenarios where the environment cannot be easily modeled [13], or the solutions are unknown a priori [14]. In this regard, recent applications derive human-level explanation based on the agent's own knowledge of the situation [15]. In particular, transforming the selected Q-values for each action into a confidence metric, using a re-shaping function based on the logarithmic transformation [15], improved the understanding of a robot trying to solve a grid-based navigation task.…”
Section: Introductionmentioning
confidence: 99%
“…In this regard, recent applications derive human-level explanation based on the agent's own knowledge of the situation [15]. In particular, transforming the selected Q-values for each action into a confidence metric, using a re-shaping function based on the logarithmic transformation [15], improved the understanding of a robot trying to solve a grid-based navigation task. The confidence metric measures how a specific action contributes to the robot reaching its goal however it does not carry any temporal correlation between the actions selected by the agent during the navigation.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation