2019
DOI: 10.1109/tnnls.2018.2885374
|View full text |Cite
|
Sign up to set email alerts
|

Plume Tracing via Model-Free Reinforcement Learning Method

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
22
0
1

Year Published

2020
2020
2023
2023

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 57 publications
(23 citation statements)
references
References 27 publications
0
22
0
1
Order By: Relevance
“…Hu et al . [100] proposed a DRL‐based method for underwater plume tracking. The RL framework follows the actor‐critic structure, using the LSTM‐based deterministic policy gradient (DPG) algorithm to learn a tracking strategy.…”
Section: Learning Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…Hu et al . [100] proposed a DRL‐based method for underwater plume tracking. The RL framework follows the actor‐critic structure, using the LSTM‐based deterministic policy gradient (DPG) algorithm to learn a tracking strategy.…”
Section: Learning Methodsmentioning
confidence: 99%
“…In Ref. [100], the reward function is adapted to the OSL background: positive rewards are gained if approaching the plume, while negative rewards are assigned if leaving the plume. Simulation results show that the proposed LSTM‐based DPG outweighs the standard DPG in efficiency and accuracy, and the resulted searching trajectory is a quite smooth one.…”
Section: Learning Methodsmentioning
confidence: 99%
“…The key underlying challenge here is to process such large amounts of IoT data and perform efficient actions. Reinforcement Learning (RL) techniques [7][8][9] are attractive solutions for this challenge (to process IoT data) as RL approaches can learn and adapt to the environmental changes and can directly learn from historical data This Project was partially supported by the Department of National Defence's Innovation for Defence Excellence and Security (IDEaS) program, Canada.…”
Section: Introductionmentioning
confidence: 99%
“…Unlike supervised learning, RL approaches do not need labeled data, which is difficult to acquire in the context of SCs. Traditionally, RL algorithms are categorized into two main classes: (i) Model-Free (MF) methods [7][8][9], which learn the value function using sample trajectories, and; (ii) Model-Based (MB) methods [12] that estimate transition and reward functions through search trees or dynamic programming. Algorithms belonging to the former category (MF algorithms), typically, fail to rapidly adjust an agent to localized changes in the reward function.…”
Section: Introductionmentioning
confidence: 99%
“…In recent years, autonomous underwater vehicles (AUVs), have undertaken important roles in oceanographic research and geoscience studies [1,2], plume tracing [3], oil spills and pipeline monitoring [4,5], underwater archaeological surveying [6], mapping flow fields [7], inspection [8], sampling [9] and tracking marine life [10] etc. In light of these autonomous applications, in this paper a solution to the issue of trajectory tracking is proposed for an under-actuated underwater vehicle.…”
Section: Introductionmentioning
confidence: 99%