2024
DOI: 10.1016/j.cels.2024.01.008
|View full text |Cite
|
Sign up to set email alerts
|

Convolutions are competitive with transformers for protein sequence pretraining

Kevin K. Yang,
Nicolo Fusi,
Alex X. Lu
Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2024
2024
2024
2024

Publication Types

Select...
5
1

Relationship

1
5

Authors

Journals

citations
Cited by 17 publications
(2 citation statements)
references
References 48 publications
0
2
0
Order By: Relevance
“…Motivated by decades of research into biophysics, molecular dynamics, and protein simulation [10, 23, 24, 27, 35], we present METL, which leverages synthetic data from molecular simulations to pretrain biophysics-aware PLMs. These biophysical pretraining signals are in contrast to existing PLMs or multiple sequence alignment-based methods that train on natural sequences and capture signals related to evolutionary selective pressures [2, 7, 8, 14, 36, 37]. By pretraining on large-scale molecular simulations, METL builds a comprehensive map of protein biophysical space.…”
Section: Discussionmentioning
confidence: 99%
“…Motivated by decades of research into biophysics, molecular dynamics, and protein simulation [10, 23, 24, 27, 35], we present METL, which leverages synthetic data from molecular simulations to pretrain biophysics-aware PLMs. These biophysical pretraining signals are in contrast to existing PLMs or multiple sequence alignment-based methods that train on natural sequences and capture signals related to evolutionary selective pressures [2, 7, 8, 14, 36, 37]. By pretraining on large-scale molecular simulations, METL builds a comprehensive map of protein biophysical space.…”
Section: Discussionmentioning
confidence: 99%
“…Some work has added graph neural network components to sequence models for downstream tasks, though these do not strictly qualify as fine-tuning methods. For example, ProtSSN 35 initializes EGNN 36 with sequence models to enhance variant prediction capabilities, MIF-ST 37 uses CARP 38 language model to boost the inverse folding task capability of graph neural networks, and ESM-GearNet 39 enhances downstream task capabilities by combining with ESM2 and GearNet.…”
Section: ■ Introductionmentioning
confidence: 99%