2021
DOI: 10.26434/chemrxiv.14416133.v1
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Transformer Neural Network for Structure Constrained Molecular Optimization

Abstract: Finding molecules with a desirable balance of multiple properties is a main challenge in drug discovery. Here, we focus on the task of molecular optimization, where a starting molecule with promising properties needs to be further optimized towards the desirable properties. Typically, chemists would apply chemical transformations to the starting molecule based on their intuition. A widely used strategy is the concept of matched molecular pairs where two molecules differ by a single transformation. In particula… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
12
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
3
1
1

Relationship

1
4

Authors

Journals

citations
Cited by 5 publications
(12 citation statements)
references
References 9 publications
0
12
0
Order By: Relevance
“…The USPTO-50K [35] dataset, which contains approximately 50 000 reactions, was used to benchmark Chemformer on the retrosynthesis prediction task. Molecular optimisation aims to improve the property profile of a starting molecule towards desirable molecular properties [6,7]. The dataset [6] for the molecular optimisation task consists of a set of matched molecular pairs (MMPs) extracted from ChEMBL [36], together with the property changes of the MMPs.…”
Section: Datasetsmentioning
confidence: 99%
See 3 more Smart Citations
“…The USPTO-50K [35] dataset, which contains approximately 50 000 reactions, was used to benchmark Chemformer on the retrosynthesis prediction task. Molecular optimisation aims to improve the property profile of a starting molecule towards desirable molecular properties [6,7]. The dataset [6] for the molecular optimisation task consists of a set of matched molecular pairs (MMPs) extracted from ChEMBL [36], together with the property changes of the MMPs.…”
Section: Datasetsmentioning
confidence: 99%
“…In particular, we examine the performance of all three pre-training tasks with base Chemformer models, after fine-tuning for 100 epochs, along with a Chemformer-Large model (pre-trained on the combined task) fine-tuned for 80 epochs. For the Transformer [6] and Transformer-R [7] benchmarks we use the published models, but examine only top-1 performance. From the table we can see that, while all Chemformer models perform strongly in comparison to existing benchmarks, the smaller Chemformer models outperform the larger on the percentage of desirable molecules generated.…”
Section: Comparison With Existing Approachesmentioning
confidence: 99%
See 2 more Smart Citations
“…Maziarka et al [37] improved self-attention with inter-atomic distances and molecular graph structure. In addition, there are some works about SMILES-based molecule generation [47,41] and optimization [18].…”
Section: Related Workmentioning
confidence: 99%