2020
DOI: 10.3390/app10228060
|View full text |Cite
|
Sign up to set email alerts
|

Online Reinforcement Learning-Based Control of an Active Suspension System Using the Actor Critic Approach

Abstract: In this paper, a controller learns to adaptively control an active suspension system using reinforcement learning without prior knowledge of the environment. The Temporal Difference (TD) advantage actor critic algorithm is used with the appropriate reward function. The actor produces the actions, and the critic criticizes the actions taken based on the new state of the system. During the training process, a simple and uniform road profile is used while maintaining constant system parameters. The controller is … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
8
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
7
1
1

Relationship

0
9

Authors

Journals

citations
Cited by 16 publications
(9 citation statements)
references
References 16 publications
1
8
0
Order By: Relevance
“…Several articles have used the reinforcement learning method, for example, that of Fares and Younes, 54 who used the critical actor algorithm, but the results obtained with the PPO algorithm are more efficient.…”
Section: Resultsmentioning
confidence: 99%
See 1 more Smart Citation
“…Several articles have used the reinforcement learning method, for example, that of Fares and Younes, 54 who used the critical actor algorithm, but the results obtained with the PPO algorithm are more efficient.…”
Section: Resultsmentioning
confidence: 99%
“…The suspension system considers the vertical movement of the body x s and that of the wheel x us along the road presented by x r . We can present the dynamics of our suspension system by the following differential equations 54 :…”
Section: Numerical Modelingmentioning
confidence: 99%
“…Our system is moving along the road that is presented by Z r . The dynamics of our suspension system are represented by the differential equations shown below 26 :…”
Section: Suspension Systemsmentioning
confidence: 99%
“…However, offline training requires a large number of labeled samples, and it is hard to label optimal control results (Ming et al., 2020). Thus, neural networks were rarely used as controllers themselves (Fares & Bani Younes, 2020).…”
Section: Introductionmentioning
confidence: 99%