2021
DOI: 10.1016/j.artint.2021.103571
|View full text |Cite
|
Sign up to set email alerts
|

Local and global explanations of agent behavior: Integrating strategy summaries with saliency maps

Abstract: With advances in reinforcement learning (RL), agents are now being developed in high-stakes application domains such as healthcare and transportation. Explaining the behavior of these agents is challenging, as the environments in which they act have large state spaces, and their decision-making can be affected by delayed rewards, making it difficult to analyze their behavior. To address this problem, several approaches have been developed. Some approaches attempt to convey the global behavior of the agent, des… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
28
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
3

Relationship

1
7

Authors

Journals

citations
Cited by 41 publications
(28 citation statements)
references
References 45 publications
(93 reference statements)
0
28
0
Order By: Relevance
“…The computational measurements done in this work present a first step to fully evaluate perturbation-based saliency maps for DRL. In the future, we will build upon the insights from this paper and conduct a human user-study, similar to the one we did in (Huber et al 2020), to evaluate how useful the saliency map approaches with good computational results are for actual end-users.…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…The computational measurements done in this work present a first step to fully evaluate perturbation-based saliency maps for DRL. In the future, we will build upon the insights from this paper and conduct a human user-study, similar to the one we did in (Huber et al 2020), to evaluate how useful the saliency map approaches with good computational results are for actual end-users.…”
Section: Discussionmentioning
confidence: 99%
“…The evaluation metrics for XAI approaches can be separated into two broad categories: human user-studies and computational measurements (Mohseni, Zarei, and Ragan 2020). Examples of human user-studies of saliency maps for DRL agents are (Huber et al 2020) and (Anderson et al 2019), who evaluate LRP and Noise Sensitivity saliency maps respectively, with regards to mental models, trust and user satisfaction. To obtain more objective quantitative data it is important to additionally evaluate explanations through computational measurements.…”
Section: Related Workmentioning
confidence: 99%
“…See Lipton (2018), Murdoch et al (2019), Rudin et al (2021) for discussions and surveys about interpretable machine learning. See some recent work on explainable RL, e.g., Amir et al (2019), Atrey et al (2020), Gottesman et al (2020), Hayes and Shah (2017), Huang et al (2019), Huber et al (2021), and Sequeira and Gervasio (2020.…”
Section: More Challengesmentioning
confidence: 99%
“…These approaches all intend to provide a global summary of the policy. Other summaries output trajectories deemed important according to importance measures [3,17] or through imitation learning [21], or train finite state representations to summarize a policy with an explainable model [7,8]. Visualization techniques combined with saliency have been used to either aggregate states and view the policy from a different perspective [49] or create a trajectory of saliency maps [14].…”
Section: Related Workmentioning
confidence: 99%
“…reduce dimension) as a function of actions [4] or formulas derived over factors of the state space [43] to output a policy summary, whereas we aggregate based on locality of the states determined by the expert policy dynamics and further identify strategic states based on these dynamics. Other summarization methods simply output simulated trajectories deemed important [3,17] as judged by whether or not the action taken at some state matters. We use the term policy dynamics to refer to state transitions and high probability paths.…”
Section: Introductionmentioning
confidence: 99%