2021
DOI: 10.3390/en14185688
|View full text |Cite
|
Sign up to set email alerts
|

Reinforcement Learning for Energy-Storage Systems in Grid-Connected Microgrids: An Investigation of Online vs. Offline Implementation

Abstract: Grid-connected microgrids consisting of renewable energy sources, battery storage, and load require an appropriate energy management system that controls the battery operation. Traditionally, the operation of the battery is optimised using 24 h of forecasted data of load demand and renewable energy sources (RES) generation using offline optimisation techniques, where the battery actions (charge/discharge/idle) are determined before the start of the day. Reinforcement Learning (RL) has recently been suggested a… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
8
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
5
1
1

Relationship

2
5

Authors

Journals

citations
Cited by 10 publications
(8 citation statements)
references
References 45 publications
0
8
0
Order By: Relevance
“…In this regard, the deviation between forecasted and actual (real) net demand by month is tracked and is 10%, 7%, −9%, −11%, 4.5%, 1.9%, −5%, −2%, −2.5%, −2%, 6.5%, and 2.3% from January to December. The convergence and cost savings of the proposed dual-layer Q-learning architecture are compared for a full year with those of the online and offline RL algorithms reported in [20]. The hyperparameters used in this work are γ, α, and ε.…”
Section: Simulation Resultsmentioning
confidence: 99%
See 1 more Smart Citation
“…In this regard, the deviation between forecasted and actual (real) net demand by month is tracked and is 10%, 7%, −9%, −11%, 4.5%, 1.9%, −5%, −2%, −2.5%, −2%, 6.5%, and 2.3% from January to December. The convergence and cost savings of the proposed dual-layer Q-learning architecture are compared for a full year with those of the online and offline RL algorithms reported in [20]. The hyperparameters used in this work are γ, α, and ε.…”
Section: Simulation Resultsmentioning
confidence: 99%
“…The solution to deal with environmental uncertainties was presented in [18,19], which uses online RL methods to find the optimal control strategy for battery operation while interacting with the real system in real time. Our previous work in [20] has compared the performance of offline RL with that of online RL for managing ESS in microgrids. Synthetic forecasted data were constructed by adding white Gaussian noise with a range of standard deviations on real data to approximate forecasted data.…”
Section: Introductionmentioning
confidence: 99%
“…These studies proposed a novel approach that integrates machine learning techniques with power management algorithms to enhance the efficiency and performance of ASD. Reinforcement learning techniques proved to be effective through simulations and experimental validations, exhibiting significant improvements in energy management and system performance Ali et al (2021).…”
Section: Discussionmentioning
confidence: 99%
“…where γ CH bat and γ D bat are binary variables representing the "on/off" and "off" states of the BESS charge and discharge, respectively. The inequality constraints for charging and discharging the BESS are shown in Equation (18).…”
Section: Economic Operation Of the Hybrid Chp System Using Milpmentioning
confidence: 99%
“…Combining CHP systems with battery storage can address these concerns; nevertheless, the dispatch of CHP and battery power must be managed to optimise the overall operation, with all generated power being consumed on-site. In an offline day-ahead optimisation, previous knowledge of the electrical load is essential, and thus, it will rely on it as forecasted load data [18]. The prediction error is likely to produce suboptimal dispatch commands that might result in power being injected into the grid, violating the 'behind-the-meter' constraints.…”
Section: Introductionmentioning
confidence: 99%