2022
DOI: 10.1093/bib/bbac037
|View full text |Cite
|
Sign up to set email alerts
|

Adapt-Kcr: a novel deep learning framework for accurate prediction of lysine crotonylation sites based on learning embedding features and attention architecture

Abstract: Protein lysine crotonylation (Kcr) is an important type of posttranslational modification that is associated with a wide range of biological processes. The identification of Kcr sites is critical to better understanding their functional mechanisms. However, the existing experimental techniques for detecting Kcr sites are cost-ineffective, to a great need for new computational methods to address this problem. We here describe Adapt-Kcr, an advanced deep learning model that utilizes adaptive embedding and is bas… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
15
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
8

Relationship

0
8

Authors

Journals

citations
Cited by 28 publications
(15 citation statements)
references
References 28 publications
0
15
0
Order By: Relevance
“…This section conducts the comparative experiments on the balanced dataset and the comparative methods including ANN [ 25 ], SVM [ 26 ], random forest (RF), K-Nearest Neighbor (KNN), FAD-BERT [ 19 ], EECL [ 10 ], Adapt_Kcr [ 40 ], and BERT4Bitter [ 20 ]. All comparative methods use one-hot to encode the amino acid residues.…”
Section: Resultsmentioning
confidence: 99%
See 1 more Smart Citation
“…This section conducts the comparative experiments on the balanced dataset and the comparative methods including ANN [ 25 ], SVM [ 26 ], random forest (RF), K-Nearest Neighbor (KNN), FAD-BERT [ 19 ], EECL [ 10 ], Adapt_Kcr [ 40 ], and BERT4Bitter [ 20 ]. All comparative methods use one-hot to encode the amino acid residues.…”
Section: Resultsmentioning
confidence: 99%
“…In this paper, we compare BERT-PPII method with the following methods. To predict PPII helices on a balanced dataset, Siermala et al ANN [25], SVM [26], random forest (RF), K-Nearest Neighbor (KNN), FAD-BERT [19], EECL [10], Adapt_Kcr [40], and BERT4Bitter [20]. All comparative methods use onehot to encode the amino acid residues.…”
Section: The Comparative Experimentsmentioning
confidence: 99%
“…In the BERT model, the input vector of each word consists of three embeddings: token embedding, segment embedding and position embedding. Position embedding can represent the absolute position information of each word in a sentence [ 47 ]. A visualization of the word embedding structure is shown in Figure 5 .…”
Section: Methodsmentioning
confidence: 99%
“…In order to fully capture the context information of the peptide, we regarded the peptide sequence as a sentence, and took k amino acids as a group, which was called “word” [ 47 ]. Since the parameter scale of BERT model is up to 100 million, it is very demanding for the experimental environment to use our own corpus to build a word list and retrain the BERT model.…”
Section: Methodsmentioning
confidence: 99%
“…Deep learning methods can automatically learn features from the raw data through convolution operations, avoiding the loss of data features. Various deep learning methods have been applied in protein sequence classification, such as bidirectional long short-term memory network (Bi-LSTM) ( Tng et al, 2021 ; Zhang Y. et al, 2022 ; Zhang et al, 2022c ; Li et al, 2022 ; Qiao et al, 2022 ; Wang et al, 2022 ), two-dimensional convolutional neural network (2D CNN) ( Le et al, 2021 ), deep residual network (ResNet) ( Xu et al, 2021 ), graph convolutional network (GCN) ( Chen et al, 2021 ), deep neural network (DNN) ( Gao et al, 2019 ; Han et al, 2019 ; Le et al, 2019 ; Hathaway et al, 2021 ), and Recurrent Neural Network (RNN) ( Zheng et al, 2020 ; Yun et al, 2021 ). These research methods have generally achieved good classification results and have attracted increasing attention.…”
Section: Introductionmentioning
confidence: 99%