2019 IEEE International Conference on Robotics and Biomimetics (ROBIO) 2019
DOI: 10.1109/robio49542.2019.8961480
|View full text |Cite
|
Sign up to set email alerts
|

Learning Adaptive Driving Behavior Using Recurrent Deterministic Policy Gradients

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
1
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
2
1

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(2 citation statements)
references
References 3 publications
0
1
0
Order By: Relevance
“…Furthermore, the implementation of the RDPG model for different applications, which require a more rapid and accurate response compared to the forecasting, verifies the suitability of the RDPG. Consequently, the RDPG was applied to control traffic lights in transportation [73], learn adaptive behavior in driving [74], perform adaptive trading for different markets [75], and to reduce errors of robot positions and joint torques [42]. These applications of the RDPG algorithm signify that the RDPG based ISVR model is feasible for both offline and real-time applications in any type of complex environment.…”
Section: Results and Comparative Analysismentioning
confidence: 99%
“…Furthermore, the implementation of the RDPG model for different applications, which require a more rapid and accurate response compared to the forecasting, verifies the suitability of the RDPG. Consequently, the RDPG was applied to control traffic lights in transportation [73], learn adaptive behavior in driving [74], perform adaptive trading for different markets [75], and to reduce errors of robot positions and joint torques [42]. These applications of the RDPG algorithm signify that the RDPG based ISVR model is feasible for both offline and real-time applications in any type of complex environment.…”
Section: Results and Comparative Analysismentioning
confidence: 99%
“…By formulating the problem as a Partially Observable Markov Decision Process (POMDP) and utilizing a Recurrent Neural Network (RNN) to consider temporal information, the agent in the RL algorithm is allowed to learn from the patterns of variance in the environment, e.g., from sensor noise, other perception limitations, or surrounding vehicle behavior, and handle the uncertainty accordingly [43,44]. To develop adaptive driving behavior, Mani et al [45] used recurrent deterministic policy gradients to train the agent how to react to the surrounding traffic condition instead of the normal DDPG. For controlling longitudinal vehicle motion specifically, Zhou et al [46] proposed a car-following model which can react to the preceding dynamics by predicting traffic oscillation using RNNs.…”
Section: B Acc Measurement Uncertaintymentioning
confidence: 99%