2020
DOI: 10.1007/978-3-030-45257-5_30
|View full text |Cite
|
Sign up to set email alerts
|

Evolutionary Context-Integrated Deep Sequence Modeling for Protein Engineering

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
6
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
6
1

Relationship

0
7

Authors

Journals

citations
Cited by 8 publications
(6 citation statements)
references
References 0 publications
0
6
0
Order By: Relevance
“…The intermediate contact map representation in the model is directly interpretable and can be used to validate the prediction or study the proteins' binding regions on a residue scale. D-SCRIPT thus joins the small but growing set of advances in interpretable deep-learning methods in computational biology (Hie et al, 2020;Luo et al, 2020aLuo et al, , 2020b. Our modular design additionally enables the investigation of model output at various stages, and we demonstrate that each layer captures incremental structural information.…”
Section: Discussionmentioning
confidence: 89%
See 1 more Smart Citation
“…The intermediate contact map representation in the model is directly interpretable and can be used to validate the prediction or study the proteins' binding regions on a residue scale. D-SCRIPT thus joins the small but growing set of advances in interpretable deep-learning methods in computational biology (Hie et al, 2020;Luo et al, 2020aLuo et al, , 2020b. Our modular design additionally enables the investigation of model output at various stages, and we demonstrate that each layer captures incremental structural information.…”
Section: Discussionmentioning
confidence: 89%
“…Chen et al's in PIPR), where each amino acid's embedding represents just its biochemical properties or a short-range context (e.g., 5-7 residues) around it. We note that alternative embeddings (Rives et al, 2019;Luo et al, 2020aLuo et al, , 2020b can potentially be substituted here.…”
Section: Model Architecturementioning
confidence: 99%
“…Protein engineering with UniRep ( 30 ) showed that general global protein representations can support training function-specific supervised models with relatively few sequence–function examples. ECNet pioneered an approach for combining global protein representations, local information about residue coevolution, and protein sequence features ( 31 ). Across tens of deep mutational scanning datasets, ECNet was almost always superior to unsupervised learning models and models based only on a global protein representation.…”
Section: Discussionmentioning
confidence: 99%
“…Bepler & Berger (2019) pre-trained LSTMs on protein sequences, adding supervision from contacts to produce embeddings. Subsequent to our preprint, related works have built on its exploration of protein sequence modeling, exploring generative models (Riesselman et al, 2019;Madani et al, 2020), internal representations of Transformers (Vig et al, 2020), and applications of representation learning and generative modeling such as classification (Elnaggar et al, 2019;Strodthoff et al, 2020), mutational effect prediction (Luo et al, 2020), and design of sequences (Repecka et al, 2019;Hawkins-Hooker et al, 2020;Amimeur et al, 2020).…”
Section: Related Workmentioning
confidence: 99%