2013
DOI: 10.1186/1471-2105-14-s17-a3
|View full text |Cite
|
Sign up to set email alerts
|

Feature selection and prediction with a Markov blanket structure learning algorithm

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
5
0

Year Published

2017
2017
2024
2024

Publication Types

Select...
2
2
1
1

Relationship

0
6

Authors

Journals

citations
Cited by 6 publications
(5 citation statements)
references
References 3 publications
0
5
0
Order By: Relevance
“…One of the most popular penalty functions is LASSO [12][13][14][15][16][17][18]. It forces most of the unimportant genes' regression coefficients into zero.…”
Section: Penalized Logistic Regression Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…One of the most popular penalty functions is LASSO [12][13][14][15][16][17][18]. It forces most of the unimportant genes' regression coefficients into zero.…”
Section: Penalized Logistic Regression Methodsmentioning
confidence: 99%
“…Filtering methods, which reduce dimensionality and try to retain the most promising features as possible, have long been under development. A number of filtering methods has been proposed to rank features, such as Information gain [13], Markov blanket [14], Bayesian variable selection [15], Boruta [16], Fisher score [17], Relief [18], maximum relevance and minimum redundancy (MRMR) [19], marginal maximum likelihood score (MMLs) [20], among which MMLS is one of the simplest and computationally efficient methods of feature selection with some criteria.…”
Section: Introductionmentioning
confidence: 99%
“…Therefore, this concept can be used to select a smaller set of relevant genes in high-dimensional problems. The MB is proven to be highly effective for feature reduction in high-dimensional problems, sometimes reducing the number of variables a thousandfold without any loss of accuracy (Aliferis, Tsamardinos, et al 2003, Shen, Li et al 2008, Fu and Desmarais 2010, Tan and Liu 2013. We put forward this idea that the MB establishment is also instrumental in creating gene module networks.…”
Section: Resultsmentioning
confidence: 98%
“…Our approach’s central basis was that protein sets which are crucial in distinguishing disease states may be key biological drivers of the disease ( 32 ). We developed a novel ML methodology that employs auxiliary Markov blanket feature selection ( 77, 78 ) combined with multiple recursive feature selection algorithms to mitigate bias towards any specific algorithm ( 79 ) and reduce overfitting, which is the fundamental challenge considering the inherent low sample size and high dimensionality of our, and many others, proteomics datasets. The first step of our method was the creation of Leave-One-Out (LOO) partitions of our data ( 35 ).…”
Section: Methodsmentioning
confidence: 99%
“…The suite of algorithms employed included RFE with Logistic Regression (LR) with L1 and L2 regularization penalties, respectively ( 30, 31 ), RFE with regularized Linear Discriminant Analysis (rLDA) ( 80 ), RFE with Random Forests (RF) ( 29 ), Boruta - Random Forests ( 81 ), and Maximum-Relevance-Minimum-Redundancy (MRMR) with an F-Statistic evaluator ( 82 ). Markov blanket feature selection was employed separately on the original datasets, due to computational expense and subsequently incorporated during the later aggregation steps ( 77, 78 ).…”
Section: Methodsmentioning
confidence: 99%