2022
DOI: 10.1016/j.epsr.2022.108546
|View full text |Cite
|
Sign up to set email alerts
|

A reinforcement learning approach to parameter selection for distributed optimal power flow

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
3
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 9 publications
(3 citation statements)
references
References 39 publications
0
3
0
Order By: Relevance
“…As the ADMM algorithm requires the optimization model to be decomposable, the auxiliary variables are used to facilitate the model decoupling (i.e., P la st = P la bt , P ev svt = P ev bvt , where P ev vt = P evc vt -P evd vt ). By incorporating these auxiliary variables and taking the negative form of ( 29), the augmented Lagrangian function of subproblem 1 can be formulated, as in (34).…”
Section: Admm Algorithm In Distributed Mannermentioning
confidence: 99%
See 1 more Smart Citation
“…As the ADMM algorithm requires the optimization model to be decomposable, the auxiliary variables are used to facilitate the model decoupling (i.e., P la st = P la bt , P ev svt = P ev bvt , where P ev vt = P evc vt -P evd vt ). By incorporating these auxiliary variables and taking the negative form of ( 29), the augmented Lagrangian function of subproblem 1 can be formulated, as in (34).…”
Section: Admm Algorithm In Distributed Mannermentioning
confidence: 99%
“…The linearization technique has been extensively explored in the existing literature [32], [33] and will not be repeated here. In addition to using the linearization technique, various machine learning techniques have been applied in the existing literature to enhance the traditional ADMM (e. g., Q-learning in [34] and deep neural networks in [35]). The incorporation of either linearization or machine learning techniques into AD-MM will not affect the effectiveness of the proposed source− load−storage low-carbon cooperative scheduling strategy.…”
Section: Admm Algorithm In Distributed Mannermentioning
confidence: 99%
“…Specifically, HMC will analyze the actions produced by a reinforcement learning agent and human, respectively, to determine which action the power system should take. Zeng et al [69] employed RL to devise an adaptive policy for selecting penalty parameters in the AC OPF problem, which is solved using the alternating direction method of multipliers. The objective of their approach is to minimize the number of iterations required for convergence.…”
Section: Learning Control Policy For Opfmentioning
confidence: 99%