2010
DOI: 10.1002/jcc.21433
|View full text |Cite
|
Sign up to set email alerts
|

Improving the accuracy of predicting disulfide connectivity by feature selection

Abstract: Disulfide bonds are primary covalent cross-links formed between two cysteine residues in the same or different protein polypeptide chains, which play important roles in the folding and stability of proteins. However, computational prediction of disulfide connectivity directly from protein primary sequences is challenging due to the nonlocal nature of disulfide bonds in the context of sequences, and the number of possible disulfide patterns grows exponentially when the number of cysteine residues increases. In … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

4
46
0

Year Published

2010
2010
2024
2024

Publication Types

Select...
9
1

Relationship

2
8

Authors

Journals

citations
Cited by 43 publications
(50 citation statements)
references
References 52 publications
4
46
0
Order By: Relevance
“…To overcome these problems, Zhu et al (2010) proposed a new feature selection technique, in which three different filter feature selection methods are used: variance score, Laplacian score and the Fisher score. Among these three, Fisher feature selection method shows the best score and selects 150 features only from the original 623 features.…”
Section: Fisher Feature Selection Methods Significantly Improves Disulmentioning
confidence: 99%
“…To overcome these problems, Zhu et al (2010) proposed a new feature selection technique, in which three different filter feature selection methods are used: variance score, Laplacian score and the Fisher score. Among these three, Fisher feature selection method shows the best score and selects 150 features only from the original 623 features.…”
Section: Fisher Feature Selection Methods Significantly Improves Disulmentioning
confidence: 99%
“…Although the increase of the feature dimension characterizes the information in the signals more completely, some irrelevant and redundant features can also cause dimension disaster [38] and negatively influence on classifier performance. In order to improve the accuracy and robustness of the TCM system, a minimum redundant feature subset needs to be obtained before the classifier is constructed.…”
Section: Feature Selectionmentioning
confidence: 99%
“…Selecting a subset of relevant features from a high-dimensional space, known as feature selection, for building robust learning models has been demonstrated to be very helpful for both improving the computational efficiency and prediction accuracy [20]. In this paper, random forest algorithm is applied to analyze the importance of each of the 2226 features.…”
Section: Feature Selectionmentioning
confidence: 99%