2017
DOI: 10.1016/j.ifacol.2017.08.747
|View full text |Cite
|
Sign up to set email alerts
|

Relations between Model Predictive Control and Reinforcement Learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
60
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
8
1

Relationship

0
9

Authors

Journals

citations
Cited by 112 publications
(60 citation statements)
references
References 46 publications
0
60
0
Order By: Relevance
“…Comparison between MPC and RL: In comparison to optimal control methods, typically employing mathematical programming, like MPC, the biggest advantage of RL is its rapid online computation. RL is model-free, adaptive, with low online complexity, but contrary to MPC, its stability and feasibility are not guaranteed and its robustness is not backed by a solid theory, while it faces difficulties in handling constraints [112]. In general, an MPC agent is expected to outperform the corresponding RL implementation, if the identified model used is accurate.…”
Section: B Reinforcement Learningmentioning
confidence: 99%
“…Comparison between MPC and RL: In comparison to optimal control methods, typically employing mathematical programming, like MPC, the biggest advantage of RL is its rapid online computation. RL is model-free, adaptive, with low online complexity, but contrary to MPC, its stability and feasibility are not guaranteed and its robustness is not backed by a solid theory, while it faces difficulties in handling constraints [112]. In general, an MPC agent is expected to outperform the corresponding RL implementation, if the identified model used is accurate.…”
Section: B Reinforcement Learningmentioning
confidence: 99%
“…Violation of such constraints may cause that the parameter update process happens outside the feasible region, which makes no sense and leads to poor transient performance. However, the constraint handling abilities of ADP methods are still immature [33], and relevant studies are still very limited as mentioned in [13], [34]. Most of existing ADP controllers neither consider the estimation bounds of unknown parameters nor can meet the prescribed performance specifications (which can also be regarded as constraints) of system states.…”
Section: Introductionmentioning
confidence: 99%
“…It can iteratively approximate the optimal control policy and avoid analytically solving the intractable HJB equation. However, the state constraint handling ability of RL-based control is still immature [23]. In this paper, a constrained RL-based control method is proposed, and a critic-only neural network (NN) structure is designed to approximate the optimal cost function and control policy.…”
Section: Introductionmentioning
confidence: 99%