2015
DOI: 10.1021/acs.jctc.5b00099
|View full text |Cite
|
Sign up to set email alerts
|

Big Data Meets Quantum Chemistry Approximations: The Δ-Machine Learning Approach

Abstract: Chemically accurate and comprehensive studies of the virtual space of all possible molecules are severely limited by the computational cost of quantum chemistry. We introduce a composite strategy that adds machine learning corrections to computationally inexpensive approximate legacy quantum methods. After training, highly accurate predictions of enthalpies, free energies, entropies, and electron correlation energies are possible, for significantly larger molecular sets than used for training. For thermochemic… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

4
875
0
10

Year Published

2015
2015
2024
2024

Publication Types

Select...
5
4

Relationship

1
8

Authors

Journals

citations
Cited by 794 publications
(930 citation statements)
references
References 48 publications
4
875
0
10
Order By: Relevance
“…The finding that the ML models also improve on the baseline method's outliers agrees with conclusions drawn in a previous finding where we applied the ∆-ML Ansatz to model DFT-level enthalpies of atomization for the 134k dataset, and where we found that for the most extreme outlier the baseline model's error reduced systematically with the training set size of the augmenting ML model. 23 When considering the 10 most extreme outliers of the ML models in Table I, neither order nor identity of the DFT outliers is conserved. Among the top 10 outliers of the 5k model, for example, there is even a saturated molecule from the opposite (blue) end of the error distribution in Fig.…”
Section: Dft and ML Model Outliersmentioning
confidence: 99%
See 1 more Smart Citation
“…The finding that the ML models also improve on the baseline method's outliers agrees with conclusions drawn in a previous finding where we applied the ∆-ML Ansatz to model DFT-level enthalpies of atomization for the 134k dataset, and where we found that for the most extreme outlier the baseline model's error reduced systematically with the training set size of the augmenting ML model. 23 When considering the 10 most extreme outliers of the ML models in Table I, neither order nor identity of the DFT outliers is conserved. Among the top 10 outliers of the 5k model, for example, there is even a saturated molecule from the opposite (blue) end of the error distribution in Fig.…”
Section: Dft and ML Model Outliersmentioning
confidence: 99%
“…20 By now, ML models have been shown to reach the highly coveted quantum chemical accuracy for many different ground-state molecular properties. [21][22][23] As such, also quantum mechanical expectation values can be interpolated in chemical space. 17 Improvement of molecular models of chemical properties based on molecular similarity 24,25 is also related to this approach.…”
Section: Introductionmentioning
confidence: 99%
“…Inspired by the high-dimensional NN reported by Behler et al, 35 we developed a neural network method named QM/MM-NN for ab initio QM/MM potential energy calculations to reduce the expensive ab initio computational cost. Similar to the Δ-machine learning method, 47 the potential energy difference between two levels was chosen as the output variable of our QM/MM-NN. We aim to develop an artificial neural network such that the NN potential energy predictions closely approximate the ab initio QM/MM calculations, and the free energy changes along the reaction coordinate through QM/MM-NN reweighting achieve similar free energy profiles.…”
Section: Introductionmentioning
confidence: 99%
“…However, the ability to learn patterns from the data increases the flexibility of machine learning algorithms. The technique is used in many fields, including determining disease risks resulting from DNA mutations,11 engineering sciences,12 space science,13 and chemistry 14…”
Section: Introductionmentioning
confidence: 99%