2019 18th IEEE International Conference on Machine Learning and Applications (ICMLA) 2019
DOI: 10.1109/icmla.2019.00145
|View full text |Cite
|
Sign up to set email alerts
|

Ensemble Bayesian Decision Making with Redundant Deep Perceptual Control Policies

Abstract: This work presents a novel ensemble of Bayesian Neural Networks (BNNs) for control of safety-critical systems. Decision making for safety-critical systems is challenging due to performance requirements with significant consequences in the event of failure. In practice, failure of such systems can be avoided by introducing redundancies of control. Neural Networks (NNs) are generally not used for safety-critical systems as they can behave in unexpected ways in response to novel inputs. In addition, there may not… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
11
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
6
3

Relationship

0
9

Authors

Journals

citations
Cited by 18 publications
(11 citation statements)
references
References 12 publications
0
11
0
Order By: Relevance
“…The racing task provides a clear objective function (fastest lap time) for the algorithm training and the race track provides with its clear driveable area and one class of objects a perfect proving ground. Researchers in this field displayed partial end-toend approaches (Lee et al, 2019;Weiss & Behl, 2020) that combine DNNs with MPC methods to create and follow dynamic trajectories. In addition, by using algorithms from the field of RL (e.g., Soft-Actor-Critic and Q-Learning), researchers were able to demonstrate how to train an agent to drive fast (de Bruin et al, 2018;Jaritz et al, 2018), how to train an agent to overtake other agents on the race track (Song et al, 2021) and how to bridge the sim-to-real gap with model-based RL approaches (Brunnbauer et al, 2021).…”
Section: Softwarementioning
confidence: 99%
“…The racing task provides a clear objective function (fastest lap time) for the algorithm training and the race track provides with its clear driveable area and one class of objects a perfect proving ground. Researchers in this field displayed partial end-toend approaches (Lee et al, 2019;Weiss & Behl, 2020) that combine DNNs with MPC methods to create and follow dynamic trajectories. In addition, by using algorithms from the field of RL (e.g., Soft-Actor-Critic and Q-Learning), researchers were able to demonstrate how to train an agent to drive fast (de Bruin et al, 2018;Jaritz et al, 2018), how to train an agent to overtake other agents on the race track (Song et al, 2021) and how to bridge the sim-to-real gap with model-based RL approaches (Brunnbauer et al, 2021).…”
Section: Softwarementioning
confidence: 99%
“…Monte-Carlo sampling methods involve sampling the parameters from a distribution and are generally obtained using an ensemble of neural network predictions. The prediction ensemble could either be generated by differently trained networks [29], or by using dropout at test-time [28].…”
Section: Epistemic Uncertainty In Graph Neural Networkmentioning
confidence: 99%
“…However, for most neural-network architectures, the true posterior distribution over parameters is intractable and must be approximated. The most common approximation scheme seen in the literature is to use dropout at test time also called MCDropout [117,196], which was successfully applied to automonous driving [130], robot control [213], and health pronostics [170]. Given a specific input x, the variance of the model's predictions from different forward passes with random shutdown of neural activations is used as a measure of the epistemic uncertainty at x.…”
Section: Epistemic Uncertaintymentioning
confidence: 99%