2022
DOI: 10.1101/2022.04.15.488492
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

A Deep SE(3)-Equivariant Model for Learning Inverse Protein Folding

Abstract: In this work, we establish a framework to tackle the inverse protein design problem; the task of predicting a protein's primary sequence given its backbone conformation. To this end, we develop a generative SE(3)-equivariant model which significantly improves upon existing autoregressive methods. Conditioned on backbone structure, and trained with our novel partial masking scheme and side-chain conformation loss, we achieve state-of-the-art native sequence recovery on structurally independent CASP13, CASP14, C… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
21
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
2
1

Relationship

1
6

Authors

Journals

citations
Cited by 16 publications
(21 citation statements)
references
References 52 publications
0
21
0
Order By: Relevance
“…Following the success of deep language models, Ferruz et al [11] developed protein sequence models to generate new proteins, but these models do not allow specification of structural motifs. Another class of methods, referred to as fixed backbone sequence design [12,20,40,27,18], attempts to solve the problem of identifying a sequence that folds into any given designable backbone structure. In the present work, we utilize a particular sequence design method, ProteinMPNN [8], but in principle any other fixed-backbone sequence design method could be used in its place.…”
Section: Appendix Table Of Contentsmentioning
confidence: 99%
“…Following the success of deep language models, Ferruz et al [11] developed protein sequence models to generate new proteins, but these models do not allow specification of structural motifs. Another class of methods, referred to as fixed backbone sequence design [12,20,40,27,18], attempts to solve the problem of identifying a sequence that folds into any given designable backbone structure. In the present work, we utilize a particular sequence design method, ProteinMPNN [8], but in principle any other fixed-backbone sequence design method could be used in its place.…”
Section: Appendix Table Of Contentsmentioning
confidence: 99%
“…Fixed backbone sequence design We use the pre-trained fixed backbone design model of (McPartlon et al, 2022) to generate sequences from backbone structures. For each test target, we generate 500 conformational decoys and sequences are deranged for each of the decoys independently.…”
Section: Loss Function and Objectivesmentioning
confidence: 99%
“…This approach comes with several drawbacks, such as the exponential search space and limited capability to model higher order interactions. Recently, many deep learning based models have explored the use of rotation equivariant frameworks (McPartlon et al, 2022; Jing et al, 2020; Hsu et al, 2022) and invaraint 3-dimensional voxel-based approaches (Qi & Zhang, 2020; Anand et al, 2022). Although our model does not directly design the sequence alongside the conformation decoys, we show that sequences designed from the resulting templates can be more reliable than those designed from a single backbone structure.…”
Section: Related Workmentioning
confidence: 99%
See 2 more Smart Citations