2023
DOI: 10.1016/j.copbio.2023.102941
|View full text |Cite
|
Sign up to set email alerts
|

Deep learning for optimization of protein expression

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
10
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 10 publications
(10 citation statements)
references
References 45 publications
0
10
0
Order By: Relevance
“…Deep learning architectures have demon-strated great predictive power across a range of sequence- to-expression tasks [19]. Convolutional neural networks (CNN), in particular, have been extensively employed to regress protein expres-sion, often achieving high performance [8, 9, 12, 13].…”
Section: Resultsmentioning
confidence: 99%
See 1 more Smart Citation
“…Deep learning architectures have demon-strated great predictive power across a range of sequence- to-expression tasks [19]. Convolutional neural networks (CNN), in particular, have been extensively employed to regress protein expres-sion, often achieving high performance [8, 9, 12, 13].…”
Section: Resultsmentioning
confidence: 99%
“…Several studies have embedded such models into algorithms for discovery of new variants using optimization methods [8] and techniques from generative models [14, 15, 16, 17]. Although the current literature has a strong focus on improvements to model architectures that can deliver greater predictive power [18], with the size of sequence- to-expression datasets growing into thousands up to millions of variants, it is becoming increasingly clear that off-the-shelf deep learning architectures such as convolutional neural networks, recurrent neural networks or transformers can readily provide high predictive accuracy [19].…”
Section: Introductionmentioning
confidence: 99%
“…Finally, most ML-based predictors represent primary protein sequences at the level of amino acids. However, synonymous mutations that do not change the amino acid sequence can still significantly impact protein expression and function , or can even relate to particular structural features. , Predictors working on the level of nucleotides or codons may thus be better tools for protein design and modification, particularly in areas such as prediction of expression, solubility, and aggregation. The fact that the 64-letter codon alphabet serves to encode richer information than the 20-letter amino acid alphabet can be directly exploited by ML models for improving performance on a wide range of tasks that are now being tackled at the protein sequence level .…”
Section: Major Gaps In the State Of The Artmentioning
confidence: 99%
“…Studies that have tackled the design of overlapping sequences, and their effects on transcriptional regulatory logic, often rely on ad hoc methods or generate a limited set of short sequences 26,[28][29][30] , thus restricting the generation of large libraries. Meanwhile, generative AI techniques are starting to show promise in emulating the complexity of contextdependent promoters [31][32][33][34][35][36][37] . However, these models often struggle with interpretability, and fine-tuning them to include or exclude specific binding sites still requires specialized expertise 38 .…”
Section: Introductionmentioning
confidence: 99%