2023
DOI: 10.1371/journal.pone.0280071
|View full text |Cite
|
Sign up to set email alerts
|

Reinforcement learning approach to control an inverted pendulum: A general framework for educational purposes

Abstract: Machine learning is often cited as a new paradigm in control theory, but is also often viewed as empirical and less intuitive for students than classical model-based methods. This is particularly the case for reinforcement learning, an approach that does not require any mathematical model to drive a system inside an unknown environment. This lack of intuition can be an obstacle to design experiments and implement this approach. Reversely there is a need to gain experience and intuition from experiments. In thi… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
3
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
7
1
1

Relationship

0
9

Authors

Journals

citations
Cited by 14 publications
(5 citation statements)
references
References 19 publications
0
3
0
Order By: Relevance
“…To encourage exploration and prevent the agent from settling into suboptimal policies, DQN integrates an ε-greedy exploration strategy. This means that the agent chooses a random action with probability ε, and chooses the action with the highest Q-value with probability 1 − ε. Israilov et al [30] applied the DQN algorithm to control the inverted pendulum on a cart both in an experimental setup and in simulation, for swing-up and stabilization of the pendulum in its unstable upward equilibrium without a dependency on initial conditions.…”
Section: Deep Q-network (Dqn)mentioning
confidence: 99%
“…To encourage exploration and prevent the agent from settling into suboptimal policies, DQN integrates an ε-greedy exploration strategy. This means that the agent chooses a random action with probability ε, and chooses the action with the highest Q-value with probability 1 − ε. Israilov et al [30] applied the DQN algorithm to control the inverted pendulum on a cart both in an experimental setup and in simulation, for swing-up and stabilization of the pendulum in its unstable upward equilibrium without a dependency on initial conditions.…”
Section: Deep Q-network (Dqn)mentioning
confidence: 99%
“…Due to the complexity of controlling a system with unstable equilibrium, an inverted pendulum (IP) is one of the most commonly used benchmark problems to test different control algorithms. A lot of studies have been conducted in the context of controlling an IP system [6], especially using RL agents [7, 8]. Most of this work focuses on algorithm development and implementation in simulation.…”
Section: Introductionmentioning
confidence: 99%
“…Bates [28] harnessed GPUs to quickly train a simulation of an inverted pendulum to balance itself. Israilov et al [29] used two model-free RL algorithms to control targets and proposed a general framework to reproduce successful experiments and simulations based on the inverted pendulum. In addition, there are still many studies of this kind, for example [30][31][32].…”
Section: Introductionmentioning
confidence: 99%