2022
DOI: 10.1016/j.jprocont.2022.06.001
|View full text |Cite
|
Sign up to set email alerts
|

Differentiable predictive control: Deep learning alternative to explicit model predictive control for unknown nonlinear systems

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
8
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
2
2

Relationship

1
7

Authors

Journals

citations
Cited by 27 publications
(8 citation statements)
references
References 54 publications
0
8
0
Order By: Relevance
“…ε−a (ν∆t + hx w + α max {a, b}). The final condition that must hold is (6), which can be checked using the following relations for y ∈ ∂C(t), x ∈ B η∆t (y): |h(x, t) − h(y, t)|= L x h x − y ≤ L x h η∆t and h(y, t) = 0 such that |h(x, t)|≤ L x h η∆t =: h, where L x h is the Lipschitz constant of h(x, t) on Ā × R. We also note that here X ⊆ R nx such that D = C ∪ Ā ⊂ X , and so the conditions of Theorem 2 hold. For α = 0.5, a = 0.03, and b = 0.00001, h is a SD-ZCBFII.…”
Section: Numerical Resultsmentioning
confidence: 99%
See 1 more Smart Citation
“…ε−a (ν∆t + hx w + α max {a, b}). The final condition that must hold is (6), which can be checked using the following relations for y ∈ ∂C(t), x ∈ B η∆t (y): |h(x, t) − h(y, t)|= L x h x − y ≤ L x h η∆t and h(y, t) = 0 such that |h(x, t)|≤ L x h η∆t =: h, where L x h is the Lipschitz constant of h(x, t) on Ā × R. We also note that here X ⊆ R nx such that D = C ∪ Ā ⊂ X , and so the conditions of Theorem 2 hold. For α = 0.5, a = 0.03, and b = 0.00001, h is a SD-ZCBFII.…”
Section: Numerical Resultsmentioning
confidence: 99%
“…The control policy is a neural network trained offline via stochastic gradient descent with gradients obtained from automatic differentiation of the MPC problem cost functions and constraints. DPC has been implemented in several applications with high performance and low computational resources [5], [6]. There exist probabilistic guarantees of safety for DPC [3], but there are no deterministic, robustness guarantees that DPC will always satisfy the constraints and stabilize the system.…”
Section: Introductionmentioning
confidence: 99%
“…The active control method exploited here [13] is an ALBC method, originally developed for dynamic control. The same approach can be applied in steady-state conditions, specifically for active wake steering.…”
Section: Hybrid Model-and Learning-based Control Methodsmentioning
confidence: 99%
“…Note that a key challenge with purely learningbased methods is that they can require a long time to train; therefore, hybrid model-and learning-based solutions, such as the solutions proposed in [13], showed to train faster than RL. To this end, hybrid methods have also shown interesting features for realworld problems, such as building control, but have not been tested on wind farm control.…”
Section: Introductionmentioning
confidence: 99%
“…In the literature the performance of MPC has been enhanced in many different ways taking advantage of machine learning techniques, by learning or improving the prediction model [26], or by obtaining explicit MPC laws to reduce the computation time [11,27]. Other works propose learning techniques to adapt the parameters of the MPC optimisation problem to improve its performance: in [15,25] bayesian optimisation is used to optimise an LQR controller (hence without enforcing constraints on states and inputs); In [5] the best linear model that approximates the nonlinear dynamics of a robotic system is chosen using bayesian optimisation in order to maximise the controller performance.…”
Section: Introductionmentioning
confidence: 99%