2021 International Conference on Software Engineering &Amp; Computer Systems and 4th International Conference on Computational 2021
DOI: 10.1109/icsecs52883.2021.00043
|View full text |Cite
|
Sign up to set email alerts
|

Comparison of PPO and SAC Algorithms Towards Decision Making Strategies for Collision Avoidance Among Multiple Autonomous Vehicles

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
7
0
1

Year Published

2022
2022
2024
2024

Publication Types

Select...
6
1

Relationship

2
5

Authors

Journals

citations
Cited by 19 publications
(8 citation statements)
references
References 19 publications
0
7
0
1
Order By: Relevance
“…Liao, et al [6] utilized the dueling deep Q-network (DDQN) to obtain a highway decision-making strategy. Muzahid, et al [30] proposed a centralized multi-vehicle control strategy by reinforcement learning (RL) and compared soft actor-critic (SAC) and proximal policy optimization (PPO) algorithms. Duan, et al [30] proposed a hierarchical structure for learning driving strategies using the RL method.…”
Section: A Decision-making Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…Liao, et al [6] utilized the dueling deep Q-network (DDQN) to obtain a highway decision-making strategy. Muzahid, et al [30] proposed a centralized multi-vehicle control strategy by reinforcement learning (RL) and compared soft actor-critic (SAC) and proximal policy optimization (PPO) algorithms. Duan, et al [30] proposed a hierarchical structure for learning driving strategies using the RL method.…”
Section: A Decision-making Methodsmentioning
confidence: 99%
“…Muzahid, et al [30] proposed a centralized multi-vehicle control strategy by reinforcement learning (RL) and compared soft actor-critic (SAC) and proximal policy optimization (PPO) algorithms. Duan, et al [30] proposed a hierarchical structure for learning driving strategies using the RL method. Chen, et al [25] built a hierarchical deep deterministic policy gradient (DDPG) algorithm and proposed an attention mechanism for learning driving strategies using images.…”
Section: A Decision-making Methodsmentioning
confidence: 99%
“…Finally, when the ego vehicle detects more than one participant in the range of the standard distance, it will make an extra careful driving situation. We need to train our model using a trial and error process to adopt our kinematic constraints 168 . Figure 7 shows the proposed conceptual framework for MVCCA in AVs, and in the following sections, we will discuss briefly all five phases.…”
Section: Conceptual Framework Of Mvccamentioning
confidence: 99%
“…In order to address the subsequent chain events of an autonomous vehicle chain collision as well as the traffic situation ontology, we will examine the problem of collision avoidance as a Markov Decision Process that can be addressed using DRL in this section. Our earlier study [43] compared two DRL methods as the foundation for selecting the DRL methodology in this work for more in-depth examination and investigation. In comparison to previous approaches such as mathematics and physics-based methods, chain collision avoidance applications using DRL do not necessitate the use of a significant mathematical model.…”
Section: ) Chain Collision Avoidance Techniquesmentioning
confidence: 99%