2021
DOI: 10.1109/tsg.2020.3047890
|View full text |Cite
|
Sign up to set email alerts
|

Deep-Reinforcement-Learning-Based Capacity Scheduling for PV-Battery Storage System

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
29
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
2
1
1

Relationship

0
9

Authors

Journals

citations
Cited by 87 publications
(29 citation statements)
references
References 17 publications
0
29
0
Order By: Relevance
“…Distributed control Proximal policy optimization 2020 [144] Capacity scheduling of PV-Battery system A proximal policy optimization-based DRL is studied to perform the capacity scheduling of PV-BSS where the method can readily leverage continuous action space and determine the specific amount of charging/discharging.…”
Section: Milpmentioning
confidence: 99%
“…Distributed control Proximal policy optimization 2020 [144] Capacity scheduling of PV-Battery system A proximal policy optimization-based DRL is studied to perform the capacity scheduling of PV-BSS where the method can readily leverage continuous action space and determine the specific amount of charging/discharging.…”
Section: Milpmentioning
confidence: 99%
“…In [18], a hybrid interval robust optimization method is proposed and the optimal SoC interval is obtained. In order to improve the profitability of photovoltaic and energy storage system investors, reference [19] uses the Proximal Policy Optimization agent to allocate system capacity. Reference [20] proposes capacity optimization that takes into account the degradation of the battery.…”
Section: B Literature Reviewmentioning
confidence: 99%
“…The objective function of the lower-level is: E is the SOC of energy storage es in the x optimization cycle. The initial and final states of stored energy are limited by (18) and (19). The state of stored energy is calculated by (20).…”
Section: ) Lower Levelmentioning
confidence: 99%
“…In addition, compared to rule-based algorithms, RL methods can learn directly from experiences without constructing an expert dataset [2], [10]. Learning-based control paradigms have been proposed for a variety of operation tasks in power grids, including control of voltage and frequency [11]- [14], capacity scheduling of PV and energy storage [15], topology control [1] and many more. A more detailed review of RL for power system operation can be found in [10].…”
Section: A Related Workmentioning
confidence: 99%