Proceedings of the 10th ACM International Conference on Bioinformatics, Computational Biology and Health Informatics 2019
DOI: 10.1145/3307339.3342135
|View full text |Cite
|
Sign up to set email alerts
|

Majority Vote Cascading

Abstract: A method to improve protein function prediction for sparsely annotated PPI networks is introduced. The method extends the DSD majority vote algorithm introduced by Cao et al. to give confidence scores on predicted labels and to use predictions of high confidence to predict the labels of other nodes in subsequent rounds. We call this a majority vote cascade. Several cascade variants are tested in a stringent cross-validation experiment on PPI networks from S. cerevisiae and D. melanogaster, and we show that for… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
2

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(1 citation statement)
references
References 25 publications
0
1
0
Order By: Relevance
“…In the most common form of k -fold cross-validation, k − 1 of the folds are used for training and the remaining fold is used for testing . Because we were especially interested in performance of function prediction methods when the amount of training data is sparse, we copied the experimental setup of Lazarsfeld et al . (2021) , which ‘inverts’ the relative sizes of the training and test data, so that, only one of the folds is used for training, and the other k − 1 have all their annotations removed and are placed in the testing set.…”
Section: Methodsmentioning
confidence: 99%
“…In the most common form of k -fold cross-validation, k − 1 of the folds are used for training and the remaining fold is used for testing . Because we were especially interested in performance of function prediction methods when the amount of training data is sparse, we copied the experimental setup of Lazarsfeld et al . (2021) , which ‘inverts’ the relative sizes of the training and test data, so that, only one of the folds is used for training, and the other k − 1 have all their annotations removed and are placed in the testing set.…”
Section: Methodsmentioning
confidence: 99%