2023
DOI: 10.1002/rnc.7083
|View full text |Cite
|
Sign up to set email alerts
|

Formation control scheme with reinforcement learning strategy for a group of multiple surface vehicles

Khai Nguyen,
Van Trong Dang,
Dinh Duong Pham
et al.

Abstract: This article presents a comprehensive approach to integrate formation tracking control and optimal control for a fleet of multiple surface vehicles (SVs), accounting for both kinematic and dynamic models of each SV agent. The proposed control framework comprises two core components: a high‐level displacement‐based formation controller and a low‐level reinforcement learning (RL)‐based optimal control strategy for individual SV agents. The high‐level formation control law, employing a modified gradient method, i… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
12
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
8

Relationship

0
8

Authors

Journals

citations
Cited by 24 publications
(12 citation statements)
references
References 54 publications
0
12
0
Order By: Relevance
“…Optimal control techniques have also been employed to maintain the formation integrity [330,331]. For instance, in [330], formation control in a leader-follower arrangement was discussed.…”
Section: Other Methodsmentioning
confidence: 99%
“…Optimal control techniques have also been employed to maintain the formation integrity [330,331]. For instance, in [330], formation control in a leader-follower arrangement was discussed.…”
Section: Other Methodsmentioning
confidence: 99%
“…Differing from typical hierarchical decision approaches, which are based on temporal or informational flow [ 20 , 21 ], the proposed non-steering/steering-based hierarchical approach in this paper aims to decrease the destabilizing risks associated with rapid steering actions. Non-steering avoidance involves trajectory adjustments through pure acceleration or deceleration without altering the vehicle’s original path, primarily focusing on longitudinal obstacle avoidance.…”
Section: Methodsmentioning
confidence: 99%
“…In Table A1, v o x (0) and v o y (0) denote the initial horizontal and longitudinal velocity of the obstacles, respectively. Once the motion of obstacles extends beyond the rectangular boundary that originated from the coordinates (20,20) to (80, 80), the subsequent speed will be set to:…”
Section: Data Availability Statementmentioning
confidence: 99%
“…The local model of a node is accepted by BCD-FL only when m t i ≥ δ, otherwise it is discarded. The federated average aggregation algorithm generally follows Equation (16).…”
Section: Model Aggregationmentioning
confidence: 99%
“…Other research initiatives, such as those detailed in [13,14], have used Stackelberg game theory and Deep Reinforcement Learning (DRL) to encourage participant engagement. This success can be attributed to the fact that reinforcement learning [15,16] has been proven to be a highly effective method in various fields of artificial intelligence. On the other hand, these solutions do not take into account the influence that the quality of model updates might have on the effectiveness of the learning process.…”
Section: Introductionmentioning
confidence: 99%