2021 IEEE Industrial Electronics and Applications Conference (IEACon) 2021
DOI: 10.1109/ieacon51066.2021.9654440
|View full text |Cite
|
Sign up to set email alerts
|

Reinforcement Learning for Cart Pole Inverted Pendulum System

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3

Citation Types

0
3
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
3
1

Relationship

0
7

Authors

Journals

citations
Cited by 8 publications
(3 citation statements)
references
References 8 publications
0
3
0
Order By: Relevance
“…More recent results are established on the data-driven technique of learning-based control. The algorithm of Deep Reinforcement Learning (DRL) allows for online extraction of the system model [9,10] or real-time identification of the optimal parameters [11,12] from the collected data. Therefore, the DRL method has significant advantages in providing an adaptive controller by which seeking optimal solutions and improving the robustness to time-varying disturbances are possible.…”
Section: Introductionmentioning
confidence: 99%
“…More recent results are established on the data-driven technique of learning-based control. The algorithm of Deep Reinforcement Learning (DRL) allows for online extraction of the system model [9,10] or real-time identification of the optimal parameters [11,12] from the collected data. Therefore, the DRL method has significant advantages in providing an adaptive controller by which seeking optimal solutions and improving the robustness to time-varying disturbances are possible.…”
Section: Introductionmentioning
confidence: 99%
“…Many numerical studies have implemented an inverted pendulum virtual environment as a benchmark to test RL algorithms [15][16][17][18][19][20][21][22], but to our knowledge, there is no study that provides successful RL implementations in experiments. First, except for a few studies that have discussed non ideal systems [16,17], most of these numerical implementations discard the effects associated to realistic (and thus more complex) control methods: in experiments, the control of the cart is subject to delay, hysteresis, biases and noise that can significantly alter the learning process.…”
Section: Introductionmentioning
confidence: 99%
“…Israilov et al [29] used two model-free RL algorithms to control targets and proposed a general framework to reproduce successful experiments and simulations based on the inverted pendulum. In addition, there are still many studies of this kind, for example [30][31][32]. However, the results of these studies focused more on reducing the time to reach equilibrium, without fully considering uncertainty in the system.…”
Section: Introductionmentioning
confidence: 99%