2022
DOI: 10.1039/d2dd00034b
|View full text |Cite
|
Sign up to set email alerts
|

Exploring chemical and conformational spaces by batch mode deep active learning

Abstract: The development of machine-learned interatomic potentials requires generating sufficiently expressive atomistic data sets. Active learning algorithms select data points on which labels, i.e., energies and forces, are calculated for inclusion...

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
22
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
6
2

Relationship

2
6

Authors

Journals

citations
Cited by 25 publications
(25 citation statements)
references
References 46 publications
(262 reference statements)
0
22
0
Order By: Relevance
“…For example, our earlier works developed ensemble-free active learning approaches for interatomic neural network (NN) potentials based on the last layer and sketched gradient features. [9][10][11] These approaches provided a learned similarity measure between data points by considering the gradient kernel of a trained NN, which corresponds to the finite-width neural tangent kernel. 12 Additionally, to avoid selecting similar structures, ref.…”
Section: Introductionmentioning
confidence: 99%
“…For example, our earlier works developed ensemble-free active learning approaches for interatomic neural network (NN) potentials based on the last layer and sketched gradient features. [9][10][11] These approaches provided a learned similarity measure between data points by considering the gradient kernel of a trained NN, which corresponds to the finite-width neural tangent kernel. 12 Additionally, to avoid selecting similar structures, ref.…”
Section: Introductionmentioning
confidence: 99%
“…AL 32,33,34 aims to iteratively collect diverse training datasets addressing any weaknesses identified in an ML model prediction. For this, it is necessary to estimate uncertainty for a model's predictions.…”
Section: Mainmentioning
confidence: 99%
“…Another feature of AL is that it can employ physically meaningful dynamical trajectories for the sampling of configurations. In the present work, we illustrate how to keep these benefits of AL, while accelerating the rate of new data collection.AL 32,33,34 aims to iteratively collect diverse training datasets addressing any weaknesses identified in an ML model prediction. For this, it is necessary to estimate uncertainty for a model's predictions.…”
mentioning
confidence: 99%
“…45 The actual measure of the uncertainty is the determinant of the corresponding MaxVol submatrix. For GM-NN, the uncertainty is computed by employing gradient features of a trained NN (with θ θ θ being the weights and biases), 26,27 corresponding to the finite-width neural tangent kernel. 46 In this work, the largest cluster maximum distance (LCMD) method for selecting new structures and last-layer gradient features (FEAT(LL)) to compute the similarity between structures and the model's uncertainty from it is used.…”
Section: Features Of the Two ML Modelsmentioning
confidence: 99%
“…ML-based models can also robustly estimate the uncertainty in their predicted energies and forces. [22][23][24][25][26][27] Such estimations are used to develop active learning algorithms to select the data most informative to the model. The data is obtained from the configurational and vibrational spaces sampled during, e.g., a molecular dynamics (MD) simulation using the existing model, either beforehand or in an on-the-fly fashion.…”
Section: Introductionmentioning
confidence: 99%