2022
DOI: 10.48550/arxiv.2203.06125
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Protein Representation Learning by Geometric Structure Pretraining

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
87
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
3

Relationship

0
8

Authors

Journals

citations
Cited by 40 publications
(87 citation statements)
references
References 0 publications
0
87
0
Order By: Relevance
“…A protein chain is essentially a sequence of amino acids, and such sequential information is shown to be crucial to determine protein functions. Hence, we follow existing studies [24,64] and integrate the sequential information in edge features. Specifically, for each edge ij that is from node i to node j, the edge feature includes an embedding of the sequential distance j − i.…”
Section: Amino Acid Levelmentioning
confidence: 99%
See 1 more Smart Citation
“…A protein chain is essentially a sequence of amino acids, and such sequential information is shown to be crucial to determine protein functions. Hence, we follow existing studies [24,64] and integrate the sequential information in edge features. Specifically, for each edge ij that is from node i to node j, the edge feature includes an embedding of the sequential distance j − i.…”
Section: Amino Acid Levelmentioning
confidence: 99%
“…Importantly, there exist hierarchical relations among different levels. Existing methods for protein representation learning either ignore hierarchical relations within proteins [26,33,20,64], or suffer from excessive computational complexity [25,21]. In this work, Figure 1: Illustration of hierarchical representations for protein structures.…”
Section: Introductionmentioning
confidence: 99%
“…Hermosilla and Ropinski [34] uses contrastive learning for representation learning of 3D protein structures from the perspective of sub-structures. Apart from that, Zhang et al [98] combines a multi-view contrastive learning and a self-prediction learning to encode geometric features of proteins. Then these semantic representations learned from SSL are utilized for downstream tasks including structure classification [35], model quality assessment [4], and function prediction [30].…”
Section: Related Workmentioning
confidence: 99%
“…They take advantage of millions of sequences to pre-train protein encoders via SSL [18,13,96]. In addition, some methods [34,98] start to directly capture the available protein structural information, which is proven to be the determinants of protein functio. Despite the fruitful progress, none of prior work considers the exploitation of temporal sequences of protein structures.…”
mentioning
confidence: 99%
“…While most of the sequence models rely on the transformer architecture, CARP [57] finds that CNNs can achieve competitive results with much fewer parameters and runtime costs. Recently, GearNet [64] explores the potential of 3D structural pre-training from the perspective of masked prediction and contrastive learning. We also summarize the molecule pretraining models in Table .1.…”
Section: Infoncementioning
confidence: 99%