2022
DOI: 10.48550/arxiv.2210.02881
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Antibody Representation Learning for Drug Discovery

Abstract: Therapeutic antibody development has become an increasingly popular approach for drug development. To date, antibody therapeutics are largely developed using large scale experimental screens of antibody libraries containing hundreds of millions of antibody sequences. The high cost and difficulty of developing therapeutic antibodies create a pressing need for computational methods to predict antibody properties and create bespoke designs. However, the relationship between antibody sequence and activity is a com… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

1
6
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
2

Relationship

2
3

Authors

Journals

citations
Cited by 5 publications
(7 citation statements)
references
References 22 publications
1
6
0
Order By: Relevance
“…We separately explored our model performance as a function of the amount of training data and demonstrated additional data, expectedly, results in improved performance [26]. However, after about 7,000 measurements, additional measurements result in less significant performance increases [26]. For this work, we trained our supervised sequence-to-affinity models on all 43,341 measurements that were available to us, but future engineering attempts may optimize use of financial resources by increasing the number of cycles while reducing the number of measurements per cycle.…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…We separately explored our model performance as a function of the amount of training data and demonstrated additional data, expectedly, results in improved performance [26]. However, after about 7,000 measurements, additional measurements result in less significant performance increases [26]. For this work, we trained our supervised sequence-to-affinity models on all 43,341 measurements that were available to us, but future engineering attempts may optimize use of financial resources by increasing the number of cycles while reducing the number of measurements per cycle.…”
Section: Discussionmentioning
confidence: 99%
“…We separately explored our model performance as a function of the amount of training data and demonstrated additional data, expectedly, results in improved performance [26]. However, after about 7,000 measurements, additional measurements result in less significant performance increases [26].…”
Section: Discussionmentioning
confidence: 99%
“…Although this paper claims finetuning AntiBERTa for paratope position prediction can achieve state-of-the-art performance, the experimental results lack standard deviations, making it unclear how significant the results obtained are. Recently, Li et al (2022) proposed a antibody-specific language model and explored its performance in SARS-CoV-2 antigen binding, showing context-dependent representations of antibody sequences benefit binding prediction.…”
Section: Pretrained Protein Language Models (Pplms)mentioning
confidence: 99%
“…Pretrained Antibody Language Models (PALMs) Encouraged by the success of PLMs in protein representation learning, series work seeks to learn antibody representations based on sequences of antibodies. (Leem et al, 2021;Ruffolo et al, 2021;Olsen et al, 2022b;Prihoda et al, 2022;Li et al, 2022). AntiBERTy (Ruffolo et al, 2021) proposed the first antibody-specific language model, exploring a Transformer trained on 558M natural antibody sequences in the OAS database.…”
Section: Pretrained Protein Language Models (Pplms)mentioning
confidence: 99%
“…Some other methods utilize antibody sequences, which are relatively cheaper and easier to obtain, to predict affinity for specific antibodies 26,27 . Although such methods employ large-scale pre-trained language models to achieve even more accurate affinity predictions 28,29 , their prediction ability is limited to the trained antigens since they do not consider the antigen information. When dealing with unknown antigens, such as in new mutation affinity tasks like XBB, the lack of antigen information is a significant limitation.…”
Section: Madani Et Al Proposed Progenmentioning
confidence: 99%