Proceedings of the 27th ACM SIGPLAN Symposium on Principles and Practice of Parallel Programming 2022
DOI: 10.1145/3503221.3508425
|View full text |Cite
|
Sign up to set email alerts
|

Extending the limit of molecular dynamics with ab initio accuracy to 10 billion atoms

Abstract: High-performance computing, together with a neural network model trained from data generated with first-principles methods, has greatly boosted applications of ab initio molecular dynamics in terms of spatial and temporal scales on modern supercomputers. Previous state-of-the-art can achieve 1 − 2 nanoseconds molecular dynamics simulation per day for 100-million atoms on the entire Summit supercomputer. In this paper, we have significantly reduced the memory footprint and computational time by a comprehensive … Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
17
0

Year Published

2022
2022
2025
2025

Publication Types

Select...
5
1

Relationship

0
6

Authors

Journals

citations
Cited by 23 publications
(19 citation statements)
references
References 39 publications
1
17
0
Order By: Relevance
“…18,19 Our implementation of the DPRc correction also supports GPU acceleration via the tensorflow library 102 and several custom operators. 60,61,103,104 The DFTB2 QM/MM+DPRc performance is only 26% slower than DFTB2 QM/MM when the correction is evaluated on a V100 NVIDIA GPU. In comparison to ab initio sampling, the DFTB2 QM/MM+DPRc method is faster by a factor of 251.…”
Section: Resultsmentioning
confidence: 99%
See 1 more Smart Citation
“…18,19 Our implementation of the DPRc correction also supports GPU acceleration via the tensorflow library 102 and several custom operators. 60,61,103,104 The DFTB2 QM/MM+DPRc performance is only 26% slower than DFTB2 QM/MM when the correction is evaluated on a V100 NVIDIA GPU. In comparison to ab initio sampling, the DFTB2 QM/MM+DPRc method is faster by a factor of 251.…”
Section: Resultsmentioning
confidence: 99%
“…The network parameters are optimized using an active learning approach described in detail elsewhere . The parameter optimizations were performed with the DP-GEN software, and the DP Compress algorithm , was applied to the trained models to improve computational performance and reduce the memory requirements. The active learning procedure involves 3 stages: training, exploration, and labeling.…”
Section: Methodsmentioning
confidence: 99%
“…1) The speed of our model exceeds both the non-reactive DeePMD benchmark on simple thermal dynamics in homogeneous systems by 30× for bulk water [26] and 4× for bulk copper [27], the SNAP benchmark by 1.7× for bulk carbon [25], and the reactive force field ReaxFF by at least 5×. ReaxFF benchmark was on 1M atoms on one V100 GPU node, and does not scale linearly with the system size.…”
Section: Innovations Realizedmentioning
confidence: 85%
“…The SNAP model was employed on bulk carbon system with peak performance 6.21 M atoms•steps/s/node [25]. Likewise, DeePMD achieved 0.3 M atoms•steps/s/node for homogeneous bulk systems -water and copper -on the OLCF Summit machine [26], with the performance recently improved to 2.0 M atoms•steps/s/node [27]. Other models, like graph neural networks, have not been shown to approach such scales due to difficulty in parallelization.…”
Section: A Scalability and Speedmentioning
confidence: 99%
See 1 more Smart Citation