2022
DOI: 10.48550/arxiv.2201.13331
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Steady-State Error Compensation in Reference Tracking and Disturbance Rejection Problems for Reinforcement Learning-Based Control

Abstract: Reinforcement learning (RL) is a promising, upcoming topic in automatic control applications. Where classical control approaches require a priori system knowledge, data-driven control approaches like RL allow a modelfree controller design procedure, rendering them emergent techniques for systems with changing plant structures and varying parameters. While it was already shown in various applications that the transient control behavior for complex systems can be sufficiently handled by RL, the challenge of non-… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2024
2024
2024
2024

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(1 citation statement)
references
References 22 publications
0
1
0
Order By: Relevance
“…16 The authors of this study also implemented steady-state error compensation as a means of enhancing the efficacy of the DDPG agent. 17 Actor-critic methods such as DDPG are in general reinforcement learning techniques in which two neural networks, named "actor" and "critic" are present. 18,19 The actor learns a policy that is modeled by a parameterized distribution, while the critic learns either a value function (i.e., a function V(s) that provides the expected return when the initial state is s and a given policy is applied) or an action-value function (i.e., a function Q(s, a) that provides the expected return obtained when the initial state is s, an arbitrary action a which may not have been obtained from the given policy is taken at s, while the given policy is used for all subsequent discrete time instants) and uses it to evaluate the performance of the policy optimized by the actor.…”
Section: Introductionmentioning
confidence: 99%
“…16 The authors of this study also implemented steady-state error compensation as a means of enhancing the efficacy of the DDPG agent. 17 Actor-critic methods such as DDPG are in general reinforcement learning techniques in which two neural networks, named "actor" and "critic" are present. 18,19 The actor learns a policy that is modeled by a parameterized distribution, while the critic learns either a value function (i.e., a function V(s) that provides the expected return when the initial state is s and a given policy is applied) or an action-value function (i.e., a function Q(s, a) that provides the expected return obtained when the initial state is s, an arbitrary action a which may not have been obtained from the given policy is taken at s, while the given policy is used for all subsequent discrete time instants) and uses it to evaluate the performance of the policy optimized by the actor.…”
Section: Introductionmentioning
confidence: 99%