2008 IEEE 2nd International Power and Energy Conference 2008
DOI: 10.1109/pecon.2008.4762658
|View full text |Cite
|
Sign up to set email alerts
|

Application of SARSA learning algorithm for reactive power control in power system

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
3
0
2

Year Published

2013
2013
2023
2023

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 8 publications
(5 citation statements)
references
References 6 publications
0
3
0
2
Order By: Relevance
“…A decentralized multiagent DRL framework that utilizes reactive power control of photovoltaics (PVs), and storage systems was proposed in paper [ 45 ]. Other RL reactive power optimizations were presented in papers [ 46 , 47 ]. Nevertheless, there is a lack of literature addressing various challenges, such as adequate formulation of the reward function and finding the optimal combination of hyperparameters that effectively leverage network architectures’ strengths and training procedures to maximize performance.…”
Section: Introductionmentioning
confidence: 99%
“…A decentralized multiagent DRL framework that utilizes reactive power control of photovoltaics (PVs), and storage systems was proposed in paper [ 45 ]. Other RL reactive power optimizations were presented in papers [ 46 , 47 ]. Nevertheless, there is a lack of literature addressing various challenges, such as adequate formulation of the reward function and finding the optimal combination of hyperparameters that effectively leverage network architectures’ strengths and training procedures to maximize performance.…”
Section: Introductionmentioning
confidence: 99%
“…Reference [14] applied DQN and Deep Deterministic Policy Gradient (DDPG) for subsystem voltage control and found that DDPG performed better with sufficient training scenarios. The voltage set point of a STATCOM is regulated using SARSA to facilitate discrete reactive power injection for voltage control in [20]. ESS, PV, and SVC output power levels are managed with the SAC algorithm to mitigate voltage violations in [21] where predefined discrete power levels are used for voltage control.…”
Section: Introductionmentioning
confidence: 99%
“…A reinforcement learning agent-in our case, a planning bot-gains decision-making knowledge by repetitively interacting with the surrounding environment (TPS) and evaluating rewards (improvement of the plan dose distribution) associated with the action (changing of optimization objectives). State-action-reward-state-action (SARSA) 7 , also known as connectionist Q-learning, is a widely-used reinforcement learning algorithm and has been proven to perform well in wide-ranging real-world applications such as controlling power systems 8 , advanced robotics 9 , and playing video games 10,11 . It is an efficient, sampling-based algorithm that sequentially changes the knowledge of the agent based on the interactive training process.…”
Section: Introductionmentioning
confidence: 99%