2022
DOI: 10.1063/5.0088019
|View full text |Cite
|
Sign up to set email alerts
|

Transfer learning using attentions across atomic systems with graph neural networks (TAAG)

Abstract: Recent advances in Graph Neural Networks (GNNs) have transformed the space of molecular and catalyst discovery. Even though the underlying physics across these domains remain the same, most prior work has focused on building domain-specific models either in small molecules or in materials. However, building large datasets across domains is computationally expensive, therefore the use of transfer learning (TL) to generalize to different domains is a promising but under-explored approach to this problem. To eval… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
31
0

Year Published

2022
2022
2025
2025

Publication Types

Select...
6
2

Relationship

4
4

Authors

Journals

citations
Cited by 27 publications
(45 citation statements)
references
References 40 publications
0
31
0
Order By: Relevance
“…For these reasons, direct approaches give reasonable performance for *OH for this use case. The best performing model is a GemNet-dT model that was initially trained to perform iterative relaxation steps, and was finetuned to directly predict the relaxed energy 25 (GemNet-dT FT). This data was used as an initial filtering step.…”
Section: Model Selectionmentioning
confidence: 99%
“…For these reasons, direct approaches give reasonable performance for *OH for this use case. The best performing model is a GemNet-dT model that was initially trained to perform iterative relaxation steps, and was finetuned to directly predict the relaxed energy 25 (GemNet-dT FT). This data was used as an initial filtering step.…”
Section: Model Selectionmentioning
confidence: 99%
“…While approaches to fine-tuning vary in which portion of the network's weights are updated, we limit our experiments to updating all the weights and leave more rigorous strategies as future work for the community. 22 For S2EF-Total , we experiment with fine-tuning using different fractions of the OC22 dataset. All fine-tuning experiments are performed using public OC20 adsorptionenergy model checkpoints found at https:// github.com/Open-Catalyst-Project/ocp/ blob/main/MODELS.md.…”
Section: Training Experimentsmentioning
confidence: 99%
“…However, there are many possible fine-tuning strategies and a large number of variations (e.g. which sections of the GNN to freeze or fit, or leaving this decision to an attention block 22 ), and we expect more progress from the community in this area. These approaches are necessary to encourage the re-use of large models, and to reduce the computational cost of obtaining stateof-the-art models for future small datasets.…”
Section: Outlook and Future Directionsmentioning
confidence: 99%
See 1 more Smart Citation
“…The scale and diversity of OC20 has additionally enabled transfer learning approaches to smaller datasets. Kolluru, et al 46 propose a transfer learning technique to use OC20 pretrained models to improve performance on smaller, out-of-distribution datasets. Similar work has also been demonstrated for other big material datasets.…”
Section: Community Progress In Developing ML Models For Catalysismentioning
confidence: 99%