2021
DOI: 10.1016/j.ast.2021.107128
|View full text |Cite
|
Sign up to set email alerts
|

A novel model-free robust saturated reinforcement learning-based controller for quadrotors guaranteeing prescribed transient and steady state performance

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
8
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
7

Relationship

1
6

Authors

Journals

citations
Cited by 24 publications
(8 citation statements)
references
References 64 publications
0
8
0
Order By: Relevance
“…In Reference 41, the actor‐NN is employed to estimate nonlinearities, actuator saturation nonlinearity and model uncertainties. The critic neural network is applied to estimate the reinforcement signals.…”
Section: Design Of Critic‐only Adp With Pgccmentioning
confidence: 99%
See 2 more Smart Citations
“…In Reference 41, the actor‐NN is employed to estimate nonlinearities, actuator saturation nonlinearity and model uncertainties. The critic neural network is applied to estimate the reinforcement signals.…”
Section: Design Of Critic‐only Adp With Pgccmentioning
confidence: 99%
“…The online actor‐NN weight update law is created to help the actor‐NN perform the estimation policy more accurately online. Different from the role of neural networks in References 41 and 42, the critic‐only NN in this paper is used to approximate the proposed PHJB and solve the optimal controller instead of estimating parameters. Moreover, a significant low computational load for the controller can be acquired because of the freedom from system dynamics obtained by adopting actor‐critic neural networks model‐free control method in References 41 and 42.…”
Section: Design Of Critic‐only Adp With Pgccmentioning
confidence: 99%
See 1 more Smart Citation
“…Remark Compared with the RL‐based controller in [48], the proposed controller in this paper employs the concept of type‐3 fuzzy logic systems in the core of the designed RL method, which leads to a more professional learning, and uses a high‐gain observer to estimate the velocities and accelerations of the QUAV system in which the velocity/acceleration sensors are removed and weight, size and cost of the QUAV system is reduced. Hence, a more efficient closed‐loop control system is achieved.…”
Section: Output Feedback Rl‐based Controller Designmentioning
confidence: 99%
“…Remarkably, more advanced RL algorithms such as soft actor-critic (SAC) [22], twin delayed deep deterministic policy gradient (TD3) [23] and proximal policy optimization (PPO) [24] are gradually being used in the control system of the quadrotor. An actor-critic neuralnetwork-based controller was presented in [25] to improve the quadrotor trajectory tracking performance. In [26], MAV has successfully completed autonomous navigation under a gust environment using the SAC algorithm as a DRL framework.…”
Section: Introductionmentioning
confidence: 99%