2022
DOI: 10.1038/s41598-022-10775-y
|View full text |Cite
|
Sign up to set email alerts
|

LM-GVP: an extensible sequence and structure informed deep learning framework for protein property prediction

Abstract: Proteins perform many essential functions in biological systems and can be successfully developed as bio-therapeutics. It is invaluable to be able to predict their properties based on a proposed sequence and structure. In this study, we developed a novel generalizable deep learning framework, LM-GVP, composed of a protein Language Model (LM) and Graph Neural Network (GNN) to leverage information from both 1D amino acid sequences and 3D structures of proteins. Our approach outperformed the state-of-the-art prot… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
35
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
2
1
1

Relationship

0
9

Authors

Journals

citations
Cited by 44 publications
(35 citation statements)
references
References 31 publications
0
35
0
Order By: Relevance
“…We hope that this work is the first step in investigating the independent and interaction effects of pretraining and architecture for protein sequence modeling. While we evaluate the effects of masked language model pretraining, transformers have also been used for autoregressive language model pretraining (Madani et al, 2020) and pairwise masked language modeling (He et al, 2021), and combining structural information (Mansoor et al, 2021; Zhang et al, 2022; McPartlon et al, 2022; Hsu et al, 2022; Chen et al, 2022; Wang et al, 2022) or functional annotations (Brandes et al, 2021) offers further directions for protein pretraining tasks.…”
Section: Discussionmentioning
confidence: 99%
“…We hope that this work is the first step in investigating the independent and interaction effects of pretraining and architecture for protein sequence modeling. While we evaluate the effects of masked language model pretraining, transformers have also been used for autoregressive language model pretraining (Madani et al, 2020) and pairwise masked language modeling (He et al, 2021), and combining structural information (Mansoor et al, 2021; Zhang et al, 2022; McPartlon et al, 2022; Hsu et al, 2022; Chen et al, 2022; Wang et al, 2022) or functional annotations (Brandes et al, 2021) offers further directions for protein pretraining tasks.…”
Section: Discussionmentioning
confidence: 99%
“…For the biological process and cellular component data sets both pre-trained methods, fine-tuned and not, outperform models trained from scratch, being the fine-tuned model the one achieving the highest F max . When compared to other methods pre-trained on large sequence data sets of millions of proteins [49,62], our method outperforms those in the molecular function data set and achieves competitive performance on the other two. When compared to the pre-trained method of Zhang et al [69] on 3D structures, our framework outperforms it on two out of three data sets.…”
Section: Protein Function Predictionmentioning
confidence: 94%
“…Results. Performance of other methods are obtained from the works of Wang et al [62] and Zhang et al [69]. For completeness, we include additional comparisons in Tbl.…”
Section: Mf Bp CCmentioning
confidence: 99%
“…While we condition on structure and reconstruct sequence, there are other methods for incorporating protein structural information, such as predicting structure similarity between protein sequences (Bepler & Berger, 2019), corrupting and reconstructing the structure in addition to the sequence (Mansoor et al, 2021; Chen et al, 2022), encoding surface features (Townshend et al, 2019), contrastive learning (Zhang et al, 2022; Cao et al, 2021), or a graph encoder without sequence decoding (Somnath et al, 2021; Fuchs et al, 2020). LM-GVP uses the same architecture as MIF-ST consisting of a pretrained language model feeding into a GNN that encodes backbone structure (Wang et al, 2022). However, in LM-GVP the structure-aware module is used as a finetuned prediction head without any pretraining.…”
Section: Related Workmentioning
confidence: 99%