2021
DOI: 10.1007/s00214-021-02725-0
|View full text |Cite
|
Sign up to set email alerts
|

Using the Gini coefficient to characterize the shape of computational chemistry error distributions

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
5
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
6
2

Relationship

3
5

Authors

Journals

citations
Cited by 9 publications
(5 citation statements)
references
References 38 publications
0
5
0
Order By: Relevance
“…The distribution of errors in density functional approximations is often not normal [5]. This can be seen in Figure 9.…”
Section: Eliminating Strange Results?mentioning
confidence: 98%
See 1 more Smart Citation
“…The distribution of errors in density functional approximations is often not normal [5]. This can be seen in Figure 9.…”
Section: Eliminating Strange Results?mentioning
confidence: 98%
“…Although these numbers convey some information, they sometimes hide the misconception that the distribution of errors is normal. In the cases we analyze below, as in most cases we studied previously [5], the distributions of errors are not normal. This justifies the use of probabilistic estimators, such as those presented in our previous work [1][2][3], or the ones introduced here.…”
Section: Statistical Measuresmentioning
confidence: 99%
“…Model errors can present any type of distribution 7 . However, a recent study on the shape of error distributions for different QoIs and DFAs showed that the distributions are most often unimodal, with various levels of asymmetry, and that they are in most cases heavier-tailed than a normal distribution 62 .…”
Section: Model Errorsmentioning
confidence: 99%
“…A contrario, it is not possible to design a prediction interval form a standard uncertainty u E i without making hypotheses on the error distribution. Being mostly dominated by model errors, computational chemistry error distributions are often non-normal, 25,26 which prevents the use of simple recipes (such as the 2σ rule). Making unsupported distribution hypotheses would add a fragility layer to the validation process, complicating the interpretation of negative validation tests.…”
Section: B Notationsmentioning
confidence: 99%
“…My goal was to derive a practical set of validation tools adapted to the specifics of CC-UQ, notably (i) the frequent use of statistical summaries (standard or expanded uncertainties), 23 instead of the prediction distributions expected by the CS framework, (ii) the possible presence of uncertainty on the reference data used for validation, (iii) the small size of most validation datasets when compared to ML applications, which limits the power of statistical validation tests, and (iv) the non-normality of the error distributions due to the frequent predominance of model errors. 9,[24][25][26] Considering these constraints, I was driven into considering two validation options, based on the available information. When expanded uncertainties are available, such as U 95 (the half-range of a 95% confidence interval), calibration should be tested by comparing the effective coverage of the corresponding prediction intervals to the target probability.…”
Section: Introductionmentioning
confidence: 99%