2017 IEEE International Conference on Robotics and Automation (ICRA) 2017
DOI: 10.1109/icra.2017.7989736
|View full text |Cite
|
Sign up to set email alerts
|

Fast discovery of influential outcomes for risk-aware MPDM

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
5
0

Year Published

2018
2018
2024
2024

Publication Types

Select...
4
1
1
1

Relationship

1
6

Authors

Journals

citations
Cited by 8 publications
(5 citation statements)
references
References 22 publications
0
5
0
Order By: Relevance
“…Model-Based. We can also cite model-based works, these methods use several models / behavior depending on the situation, such as Multi-Policy Decision-Making [23,24] which uses the policy with the best expected utility among a set of different policies. Sebastian et al [25] proposed a Gaussian Mixture Model in order to differentiate people's behavior and choose the most relevant and respectful path towards people.…”
Section: Supervising Learningmentioning
confidence: 99%
“…Model-Based. We can also cite model-based works, these methods use several models / behavior depending on the situation, such as Multi-Policy Decision-Making [23,24] which uses the policy with the best expected utility among a set of different policies. Sebastian et al [25] proposed a Gaussian Mixture Model in order to differentiate people's behavior and choose the most relevant and respectful path towards people.…”
Section: Supervising Learningmentioning
confidence: 99%
“…In each Monte Carlo sample, all the agents are forward simulated together to capture closed-loop interactions [29] and the best ego agent policy is elected from the cumulative results. Further work in the MPDM framework focuses on discovering risky configurations through stochastic gradient descent of a heuristic cost function [30] and back-propagation techniques [31].…”
Section: F Multi-policy Approachesmentioning
confidence: 99%
“…Moreover, the hidden intentions (driving policy of other agents) are sampled according to initial behavioral prediction (initial belief) and will not be updated during the simulation. As a result, risky outcomes may not be reflected in policy evaluation due to inaccurate initial behavior prediction or insufficient intention samples [17].…”
Section: Related Workmentioning
confidence: 99%
“…In the case of MPDM, the intention of the nearby vehicles is fixed for the whole planning horizon, and the initial intention is sampled according to a behavior prediction algorithm. The limitation of MPDM is that, with a limited number of samples, influential risky outcomes may not be rolled out, especially when the initial intention prediction is inaccurate [17].…”
Section: Conditional Focused Branchingmentioning
confidence: 99%
See 1 more Smart Citation