2022
DOI: 10.1101/2022.02.03.479055
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Towards Effective and Generalizable Fine-tuning for Pre-trained Molecular Graph Models

Abstract: Graph Neural Networks (GNNs) and Transformer have emerged as dominant tools for AI-driven drug discovery. Many state-of-the-art methods first pre-train GNNs or the hybrid of GNNs and Transformer on a large molecular database and then fine-tune on downstream tasks. However, different from other domains such as computer vision (CV) or natural language processing (NLP), getting labels for molecular data of downstream tasks often requires resource-intensive wet-lab experiments. Besides, the pre-trained m… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
9
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
2
1

Relationship

2
5

Authors

Journals

citations
Cited by 9 publications
(9 citation statements)
references
References 18 publications
0
9
0
Order By: Relevance
“…(Zhang et al 2022) adapts the optimal transport to constrain the fine-tuned model behaviors, which is a kind of representation regularization. (Xia et al 2022) uses a regularization built on dropout to control the complexity of pre-trained models. Although there are various forms of fine-tuning, it is evident that a gap exists between the learning objectives of the pre-training task and the downstream task.…”
Section: Preliminariesmentioning
confidence: 99%
See 1 more Smart Citation
“…(Zhang et al 2022) adapts the optimal transport to constrain the fine-tuned model behaviors, which is a kind of representation regularization. (Xia et al 2022) uses a regularization built on dropout to control the complexity of pre-trained models. Although there are various forms of fine-tuning, it is evident that a gap exists between the learning objectives of the pre-training task and the downstream task.…”
Section: Preliminariesmentioning
confidence: 99%
“…Though the performance is promising, the use of optimal transport requires a high computational cost, thus not applicable for large-scale models. As another example, Xia et al (2022) focuses on molecular graphs, and proposes a new regularization tailored to pre-trained molecular model, but this approach can only be applied to molecular graphs. As a newly developed research direction, prompt-tuning has attracted considerable attention recently.…”
Section: Related Workmentioning
confidence: 99%
“…We plot the training and validation curves in Figure 7, from which we can observe that the pre-trained models outperform training from scratch by significant margins. Additionally, for small-scale datasets such as Bace, training from scratch tends to overfit the training data of downstream tasks (Xia et al, 2022d). In contrast, the pre-trained models can mitigate the over-fitting issue.…”
Section: E Implementation Details Of Baselinesmentioning
confidence: 99%
“…To evaluate the transferability of the pre-training scheme, we conduct experiments on transfer learning on molecular property prediction in chemistry and protein function prediction in biology following previous works [16,49,54]. Specifically, we pre-train and finetune the models with different datasets.…”
Section: Transferability (Rq2)mentioning
confidence: 99%