2022
DOI: 10.1016/j.neunet.2022.03.022
|View full text |Cite
|
Sign up to set email alerts
|

MoËT: Mixture of Expert Trees and its application to verifiable reinforcement learning

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
7
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
2
2
1

Relationship

0
9

Authors

Journals

citations
Cited by 19 publications
(7 citation statements)
references
References 25 publications
0
7
0
Order By: Relevance
“…During the learning process, this clustering became more distinct, and in the 'expert' agent the clusters were associated with specific phases of θ during the circling motion around the thermal center. Further analyzing these clusters and their evolution during the learning process may inform us about the inner workings of the NN and may be used to understand the agent's policy by extracting explicit rules, for example, by describing each cluster as a decision tree [39][40][41][42] . In summary, the application of deep-RL method as presented here may contribute to further improving autonomous UAV gliding systems and to our understanding of motion control learning.…”
Section: Discussionmentioning
confidence: 99%
“…During the learning process, this clustering became more distinct, and in the 'expert' agent the clusters were associated with specific phases of θ during the circling motion around the thermal center. Further analyzing these clusters and their evolution during the learning process may inform us about the inner workings of the NN and may be used to understand the agent's policy by extracting explicit rules, for example, by describing each cluster as a decision tree [39][40][41][42] . In summary, the application of deep-RL method as presented here may contribute to further improving autonomous UAV gliding systems and to our understanding of motion control learning.…”
Section: Discussionmentioning
confidence: 99%
“…The mechanistic interpretability of a standard deep-learning approach (cf. [281][282][283][284][285][286][287][288][289][290][291][292]) is limited with nearly a black box and one typically not amenable to the individual differences here given model demands for data and dimensionality that are orders of magnitude larger. Hence, despite general merits of deep learning, the promise here is confronted by formidable challenges both practical and epistemological.…”
Section: Dynamics Of Hysteresismentioning
confidence: 99%
“…With analogies between animal learning [6][7][8][55][56][57][380][381][382] and machine learning [49][50][51][52][53][54]288,291,[383][384][385][386][387][388][389][390][391][392][393][394][395][396], the theory of a mixture of experts is based on advantages of modular parallelism and conditional computation for balancing versatility and efficiency in optimal control. As with the mixture-of-experts (MoE) architecture per se (which has also proven effective for sparse scaling of a deep neural network), the scope of this consilient theory can be extended to systems of varying levels of expertise as well as nonexpert controllers and their numerous choice and action biases (cf.…”
Section: The Optimality Of Nonexpert Control With Lessons For ML and Aimentioning
confidence: 99%
“…These include approaches based on SMT-solving [15], [27], [30]- [32], [35], approaches based on reachability analysis and abstract interpretation [21], [40], [56], [63], and many others. Recently, the verification of DRL-trained DNNs and the systems in which they operate also received some attention [5], [9], [19], [33], [65]. Our work here is another step toward applying DNN verification techniques to real systems of interest.…”
Section: Related Workmentioning
confidence: 99%