2019
DOI: 10.1007/978-3-030-37442-6_3
|View full text |Cite
|
Sign up to set email alerts
|

Automatic Collision Avoidance Using Deep Reinforcement Learning with Grid Sensor

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
8
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
5

Relationship

1
4

Authors

Journals

citations
Cited by 5 publications
(8 citation statements)
references
References 8 publications
0
8
0
Order By: Relevance
“…The policy and value function used in PPO are represented by deep neural networks. In the present study, we set a safe distance of 0.5 NM in the discrete action space in advance, but the trained model by previous approach [17] did not reach the sufficient performance. One of the possible reasons is that networks consist of only convolutional layers and fullconnected layers (FC) cannot store historical information of the environment.…”
Section: Structure Of Network and Update Methodsmentioning
confidence: 95%
See 4 more Smart Citations
“…The policy and value function used in PPO are represented by deep neural networks. In the present study, we set a safe distance of 0.5 NM in the discrete action space in advance, but the trained model by previous approach [17] did not reach the sufficient performance. One of the possible reasons is that networks consist of only convolutional layers and fullconnected layers (FC) cannot store historical information of the environment.…”
Section: Structure Of Network and Update Methodsmentioning
confidence: 95%
“…The hyperparameters for PPO in continuous action spaces are provided in Table 3. The hyperparameters of the previous model in discrete action spaces are described in [17].…”
Section: Structure Of Network and Update Methodsmentioning
confidence: 99%
See 3 more Smart Citations