2020
DOI: 10.1109/access.2020.3015661
|View full text |Cite
|
Sign up to set email alerts
|

Navigation in Restricted Channels Under Environmental Conditions: Fast-Time Simulation by Asynchronous Deep Reinforcement Learning

Abstract: This paper proposes an efficient method, based on reinforcement learning, to be used as ship controller in fast-time simulators within restricted channels. The controller must operate the rudder in a realistic manner in both time and angle variation so as to approximate human piloting. The method is well suited to scenarios where no previous navigation data is available; it takes into account, during training, both the effect of environmental conditions and also curves in channels. We resort to an asynchronous… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
3
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
7
3

Relationship

0
10

Authors

Journals

citations
Cited by 16 publications
(5 citation statements)
references
References 18 publications
0
3
0
Order By: Relevance
“…Among these techniques, Deep learning-based RL architectures like deep Q-Network [24], deep deterministic policy gradient [25], PPO [26], and Soft Actor-Critic [27] have demonstrated exceptional capabilities in handling complex stochastic tasks. These methods have proven to be effective in areas such as cooperative path planning [28], navigating constrained environments [29], and patrolling water resources [30]. However, challenges arise in training tasks that require prolonged processes due to the inherent behavior of RL agents that rely on a specific, dense, and short-term reward system [31], [32].…”
Section: E Reinforcement Learningmentioning
confidence: 99%
“…Among these techniques, Deep learning-based RL architectures like deep Q-Network [24], deep deterministic policy gradient [25], PPO [26], and Soft Actor-Critic [27] have demonstrated exceptional capabilities in handling complex stochastic tasks. These methods have proven to be effective in areas such as cooperative path planning [28], navigating constrained environments [29], and patrolling water resources [30]. However, challenges arise in training tasks that require prolonged processes due to the inherent behavior of RL agents that rely on a specific, dense, and short-term reward system [31], [32].…”
Section: E Reinforcement Learningmentioning
confidence: 99%
“…These research also have only considered the constant radius of safety region to the obstacle. Trajectory planning on other kinds of restricted waters have been reported on: global trajectory planning and collision avoidance of an autonomous surface vessel at narrow ferry passage by safety region and collision region around moving and static obstacles with applying the hybrid dynamic window method (Serigstad et al, 2018); Navigation in restricted channels using reinforcement learning was done by Amendola et al (2020).…”
Section: Related Researchmentioning
confidence: 99%
“…Besides those focused on berthing/unberthing maneuver, trajectory/path planning methods exist for generating reference trajectories for automation of ship operation. As an application of reinforcement learning, Amendola et al (2020) investigated the acquisition of control laws for narrow channel maneuver. Their method provided solutions that could be used as a trajectory plan.…”
Section: Introductionmentioning
confidence: 99%