2022
DOI: 10.1007/s10489-022-03605-1
|View full text |Cite
|
Sign up to set email alerts
|

Explaining deep reinforcement learning decisions in complex multiagent settings: towards enabling automation in air traffic flow management

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
4
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
2
2
2

Relationship

0
6

Authors

Journals

citations
Cited by 7 publications
(4 citation statements)
references
References 40 publications
0
4
0
Order By: Relevance
“…Given a minibatch of states, we calculate the MAE of this minibatch for any action as the mean absolute difference between the Q-values estimated by the mimic learner and the Q-values estimated by the deep Q-network for that action. More formally, for a minibatch of states 𝐷 𝑠 , the MAE 𝑖 of action 𝑎 𝑖 is denoted as: We focus on providing aggregated interpretations, focusing on the contribution of features to local decisions and to the overall policy: This, as suggested by ATM operators, is beneficial towards understanding decisions, helping them to increase their confidence to the solutions proposed, and mastering the inherent complexity in such a multi-agent setting, as solutions may be due to complex phenomena that are hard to be traced [15]. Specifically, in this work, local explainability measures state features' importance on a specific instance (i.e.…”
Section: Evaluation Metrics and Methodsmentioning
confidence: 99%
See 3 more Smart Citations
“…Given a minibatch of states, we calculate the MAE of this minibatch for any action as the mean absolute difference between the Q-values estimated by the mimic learner and the Q-values estimated by the deep Q-network for that action. More formally, for a minibatch of states 𝐷 𝑠 , the MAE 𝑖 of action 𝑎 𝑖 is denoted as: We focus on providing aggregated interpretations, focusing on the contribution of features to local decisions and to the overall policy: This, as suggested by ATM operators, is beneficial towards understanding decisions, helping them to increase their confidence to the solutions proposed, and mastering the inherent complexity in such a multi-agent setting, as solutions may be due to complex phenomena that are hard to be traced [15]. Specifically, in this work, local explainability measures state features' importance on a specific instance (i.e.…”
Section: Evaluation Metrics and Methodsmentioning
confidence: 99%
“…The tuple containing all agents' local states is the joint global state. Q-learning [33] agents has been shown to achieve remarkable performance on this task [15]. In our experiments, all agents share parameters and replay buffer and act independently.…”
Section: Real-world Demand-capacity Problem Settingmentioning
confidence: 95%
See 2 more Smart Citations