2023
DOI: 10.1186/s13321-023-00700-4
|View full text |Cite
|
Sign up to set email alerts
|

Double-head transformer neural network for molecular property prediction

Abstract: Existing molecular property prediction methods based on deep learning ignore the generalization ability of the nonlinear representation of molecular features and the reasonable assignment of weights of molecular features, making it difficult to further improve the accuracy of molecular property prediction. To solve the above problems, an end-to-end double-head transformer neural network (DHTNN) is proposed in this paper for high-precision molecular property prediction. For the data distribution characteristics… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
4
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 8 publications
(4 citation statements)
references
References 59 publications
0
4
0
Order By: Relevance
“…Many of the constructed ADMET Transformer models make use of those popularized in NLP literature, such as but not limited to BERT, RoBERTa, and GPT-2 [32][33][34][35][36][37][38][39][40][41][42]. Others combine graph representations with the Transformer to obtain graph-level contextual understanding [43][44][45][46][47][48][49]. In addition, some works use a combination of molecular line notation and pre-fabricated descriptors [38], while the remaining use the Transformer with various training strategies and architectural changes [50][51][52][53].…”
Section: Transformer-based Admet Modelsmentioning
confidence: 99%
“…Many of the constructed ADMET Transformer models make use of those popularized in NLP literature, such as but not limited to BERT, RoBERTa, and GPT-2 [32][33][34][35][36][37][38][39][40][41][42]. Others combine graph representations with the Transformer to obtain graph-level contextual understanding [43][44][45][46][47][48][49]. In addition, some works use a combination of molecular line notation and pre-fabricated descriptors [38], while the remaining use the Transformer with various training strategies and architectural changes [50][51][52][53].…”
Section: Transformer-based Admet Modelsmentioning
confidence: 99%
“…This generative design approach successfully reproduced highly potent compounds for different activity classes based on weakly potent input compounds 13 . Transformer models have also been derived for other compound property predictions [14][15][16] and generative compound design applications [17][18][19] as well as for the prediction of drug-target interactions [20][21][22] .…”
Section: Meta-learning For Transformer-based Prediction Of Potent Com...mentioning
confidence: 99%
“…TranGRU , on the other hand, enhances the understanding of both local and global molecular information, positioning itself as a versatile sequence encoder for molecular representation extraction. DHTNN , a novel algorithmic framework, introduces the innovative Beaf activation function and leverages a Transformer with Double-head attention for molecular feature extraction, resulting in a robust approach that ensures model convergence and rational weight assignments [ 149 ]. Two strategies, MolHGT and PharmHGT , both of them applied the Heterogeneous Graph Transformer mechanism in molecular property research.…”
Section: Applications Of Attention-based Models In Drug Discoverymentioning
confidence: 99%