2016
DOI: 10.2166/hydro.2016.243
|View full text |Cite
|
Sign up to set email alerts
|

A novel nested stochastic dynamic programming (nSDP) and nested reinforcement learning (nRL) algorithm for multipurpose reservoir optimization

Abstract: In this article we present two novel multipurpose reservoir optimization algorithms named nested stochastic dynamic programming (nSDP) and nested reinforcement learning (nRL). Both algorithms are built as a combination of two algorithms; in the nSDP case it is (1) stochastic dynamic programming (SDP) and (2) nested optimal allocation algorithm (nOAA) and in the nRL case it is (1) reinforcement learning (RL) and (2) nOAA. The nOAA is implemented with linear and non-linear optimization. The main novel idea is to… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
6
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
1

Relationship

0
6

Authors

Journals

citations
Cited by 9 publications
(6 citation statements)
references
References 12 publications
0
6
0
Order By: Relevance
“…The analysis of the reservoir network resulted in a schematization of the system, which was the base for the operational model (Fig. 3 (Frank et al, 2016) with the M5 Rules classifier. For more information on the data-driven model and its accuracy, see Appendix S1.…”
Section: Operational Modelmentioning
confidence: 99%
See 2 more Smart Citations
“…The analysis of the reservoir network resulted in a schematization of the system, which was the base for the operational model (Fig. 3 (Frank et al, 2016) with the M5 Rules classifier. For more information on the data-driven model and its accuracy, see Appendix S1.…”
Section: Operational Modelmentioning
confidence: 99%
“…The daily inflow time series are freely available for 2002-2015 from XM Compañía de Expertos en Mercados (2017), except INPorce3, which is available from 2011, the year when Porce III was put into operation. Therefore, INPorce3 has been extended to 2002-2015 with a validated data-driven model of high accuracy (RMSE 18.4 m 3 /s, correlation coefficient 0.9869) that has been created using the Weka Workbench software (Frank et al, 2016) with the M5 Rules classifier. For more information on the data-driven model and its accuracy, see Appendix S1.…”
Section: Operational Modelmentioning
confidence: 99%
See 1 more Smart Citation
“…Classical tabular RL is closely related to dynamic programming and has been explored for multiobjective reservoir management. [22][23][24][25][26] However, because tabular RL is limited to systems with relatively small numbers of possible states and actions, deep reinforcement learning (also widely referred to as RL), which uses neural networks as function approximators instead of using lookup tables, has been used for control of more complex systems. 27,28 This approach to learning allows RL increased flexibility to optimize control actions, balance competing objectives based on the formulation of the reward function, and has the potential to continually adapt system controls to evolving environmental conditions (e.g., increased runoff from urbanization or climate change).…”
Section: Introductionmentioning
confidence: 99%
“…Different methods have been examined to evaluate the effectiveness of re-operating or optimizing the operation of reservoirs in order to maximize the benefits, especially hydropower generation, in particular using optimization methods e.g., [15][16][17][18]. Researchers have tested classic optimization approaches such as linear programming (LP) [19,20], non-linear programming (NLP) [21,22], and dynamic programming (DP) [23,24] to solve reservoir operation problems in different forms. However, these studies also simplified objective functions and constraints in a linear model, which makes them unsuitable for optimizing hydropower generation [25].…”
Section: Introductionmentioning
confidence: 99%