2021 IEEE International Conference on Robotics and Automation (ICRA) 2021
DOI: 10.1109/icra48506.2021.9560962
|View full text |Cite
|
Sign up to set email alerts
|

Risk-Conditioned Distributional Soft Actor-Critic for Risk-Sensitive Navigation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

0
4
0

Year Published

2021
2021
2025
2025

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 17 publications
(4 citation statements)
references
References 26 publications
0
4
0
Order By: Relevance
“…Risk-sensitive planning and control dates back to the 1970s, as exemplified by risk-sensitive Linear-Exponential-Quadratic-Gaussian [49,50] and risk-sensitive Markov Decision Processes (MDPs) [51]. More recent methods include risk-sensitive nonlinear MPC [6,52], Q-learning [44,53], and actor-critic [54,55] methods, for various types of risk measures. Refer to a recent survey [56] for further details.…”
Section: Related Workmentioning
confidence: 99%
“…Risk-sensitive planning and control dates back to the 1970s, as exemplified by risk-sensitive Linear-Exponential-Quadratic-Gaussian [49,50] and risk-sensitive Markov Decision Processes (MDPs) [51]. More recent methods include risk-sensitive nonlinear MPC [6,52], Q-learning [44,53], and actor-critic [54,55] methods, for various types of risk measures. Refer to a recent survey [56] for further details.…”
Section: Related Workmentioning
confidence: 99%
“…We aim to improve over both by providing a dynamic risk-adjustment method that is sensitive to change in both the policy and environment and thus broadly applicable. Choi et al [2021] employ a dynamic approach by conditioning the policy on the risk level α. Their method requires policy evaluation and training across different uniformly distributed α levels, which dramatically increases the number of observations and therefore the computational complexity needed for training.…”
Section: Related Workmentioning
confidence: 99%
“…The idea of ARA is based on the parametric uncertainty of the agent's predictions to adapt its risk-awareness, thus encouraging more cautious behaviour in states that have not been visited as often. As an additional advantage over prior work [Choi et al, 2021], it does not add significant overhead to training as it requires no additional episode evaluations.…”
Section: Introductionmentioning
confidence: 99%
“…Compared to DRL, Distributional RL methods capture return distributions instead of only the expected return [12], and these methods have been used to train risk-sensitive agents for a variety of navigation tasks [34], [35], [36], [37], [14] in recent years. These works focus on the single robot navigation problem in environments with static obstacles or dynamic obstacles moving along pre-defined trajectories.…”
Section: Related Workmentioning
confidence: 99%