2014
DOI: 10.1016/j.envsoft.2014.02.013
|View full text |Cite
|
Sign up to set email alerts
|

Evaluating, interpreting, and communicating performance of hydrologic/water quality models considering intended use: A review and recommendations

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

2
91
0
5

Year Published

2015
2015
2022
2022

Publication Types

Select...
9

Relationship

0
9

Authors

Journals

citations
Cited by 132 publications
(98 citation statements)
references
References 46 publications
2
91
0
5
Order By: Relevance
“…These include multicomponent mapping (Pappenberger and Beven, 2004), self-organizing maps (Reusser et al, 2009), wavelets (Liu et al, 2011), the hydrograph matching algorithm (Ewen, 2011), and the "Peak-Box" approach for the interpretation and verification of operational ensemble peakflow forecasts (Zappa et al, 2013). Despite this considerable progress, many practical and scientific applications (Haag et al, 2005;Gassmann et al, 2013;Seibert et al, 2014;Wrede et al, 2014;Kelleher et al, 2015;Zhang et al, 2016) still rely on simple mean squared error (MSE) type distance metrics such as the long-established Nash-Sutcliffe efficiency (NASH) or the root mean squared error (RMSE) even though their shortcomings are well known (Seibert, 2001;Schaefli and Gupta, 2007;Gupta et al, 2009). A less recognized issue of MSE-type criteria is that these compare points with identical abscissa, i.e.…”
Section: Single and Multiple Criteria For Hydrograph Evaluationmentioning
confidence: 99%
See 1 more Smart Citation
“…These include multicomponent mapping (Pappenberger and Beven, 2004), self-organizing maps (Reusser et al, 2009), wavelets (Liu et al, 2011), the hydrograph matching algorithm (Ewen, 2011), and the "Peak-Box" approach for the interpretation and verification of operational ensemble peakflow forecasts (Zappa et al, 2013). Despite this considerable progress, many practical and scientific applications (Haag et al, 2005;Gassmann et al, 2013;Seibert et al, 2014;Wrede et al, 2014;Kelleher et al, 2015;Zhang et al, 2016) still rely on simple mean squared error (MSE) type distance metrics such as the long-established Nash-Sutcliffe efficiency (NASH) or the root mean squared error (RMSE) even though their shortcomings are well known (Seibert, 2001;Schaefli and Gupta, 2007;Gupta et al, 2009). A less recognized issue of MSE-type criteria is that these compare points with identical abscissa, i.e.…”
Section: Single and Multiple Criteria For Hydrograph Evaluationmentioning
confidence: 99%
“…For this reason different attempts have been undertaken to compare expert judgement and automated criteria (Crochemore et al, 2014) and to establish model evaluation guidelines (e.g. Moriasi et al, 2007;Biondi et al, 2012;Harmel et al, 2014). Key points of related guidelines typically include the statement that the choice of the metric should depend (i) on the modelling purpose, (ii) on the modelling mode (calibration, validation, simulation, or forecast), and (iii) on the model resolution (time stepping, spatial resolution).…”
Section: Single and Multiple Criteria For Hydrograph Evaluationmentioning
confidence: 99%
“…• Most studies did not report any estimate of measurement uncertainty (i.e., uncertainty due to an instrument or method), even though researchers are increasingly calling for and explaining the benefits of publishing such values corresponding to measured data (Beven, 2006;Harmel et al, 2006aHarmel et al, , 2009Harmel et al, , 2014). …”
Section: Challenges and Issues In Systematic Reviews Syntheses And mentioning
confidence: 99%
“…We recommend comparing observed and simulated time series data in two ways (e.g., van Werkhoven et al, 2009;Hrachowitz et al, 2014): as statistical metrics, which measure model performance with respect to the entire time series, e.g., root mean squared error (RMSE), Nash-Sutcliffe efficiency coefficient (NSE; Nash and Sutcliffe, 1970), and dynamic metrics, which measure model performance with respect to different periods or types of hydrologic behavior, e.g., the baseflow index, the slope of the flow duration curve (Wagener et al, 2001;Gupta et al, 2008;Pfannerstill et al, 2014;Shafii and Tolson, 2015). Performance across statistical metrics is typically judged with respect to a threshold value, e.g., NSE greater than 0.8, or some threshold percentage, e.g., top 10 % of RMSE values (e.g., Moriasi et al, 2007;Harmel et al, 2014). Dynamic metrics may expand assessment of hydrologic behavior, as existing work has shown that there is information contained not only in different types of data but also in different periods for an observational time series (Wagener et al, 2001;Gupta et al, 2008;Pfannerstill et al, 2014;Shafii and Tolson, 2015).…”
Section: Frameworkmentioning
confidence: 99%