2019
DOI: 10.48550/arxiv.1907.01180
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Conservative Q-Improvement: Reinforcement Learning for an Interpretable Decision-Tree Policy

Abstract: There is a growing desire in the field of reinforcement learning (and machine learning in general) to move from black-box models toward more "interpretable AI." We improve interpretability of reinforcement learning by increasing the utility of decision tree policies learned via reinforcement learning. These policies consist of a decision tree over the state space, which requires fewer parameters to express than traditional policy representations. Existing methods for creating decision tree policies via reinfor… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
15
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
4
3
2

Relationship

1
8

Authors

Journals

citations
Cited by 13 publications
(15 citation statements)
references
References 9 publications
0
15
0
Order By: Relevance
“…Several works [5,6,7,8] use greedy heuristics to induce the trees. However, this approaches have the following drawbacks:…”
Section: Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…Several works [5,6,7,8] use greedy heuristics to induce the trees. However, this approaches have the following drawbacks:…”
Section: Methodsmentioning
confidence: 99%
“…In [5] the authors propose a method that predicts the gain obtained by adding a split to the tree and select the best split to grow the tree. The experimental results show that this method is more effective than the method proposed in [8] on the tested environment.…”
Section: Theorymentioning
confidence: 99%
“…The authors show that their proposition can achieve comparable performance to the original non-interpretable policy, and is amenable to verification. As an alternative approach to control the decision tree size, [200] proposes to increase its size only if the novel decision tree increases sufficiently the performance. As an improvement to work (like VIPER) using only one decision tree, In [201], a mixture of Expert Trees (MOET) is proposed.…”
Section: Decision Trees and Variantsmentioning
confidence: 99%
“…Other explanation methods include decision tree (Bastani et al, 2018;Gupta et al, 2015;Roth et al, 2019) and structural causal MDP (Madumal et al, 2019;Waa et al, 2018). While decision tree can be represented graphically and thus aid in human understanding, a reasonably-sized tree with explainable attributes is difficult to construct, especially in the vision-based domain.…”
Section: Explaining Traditional Rl Agentsmentioning
confidence: 99%