2016
DOI: 10.1515/sagmb-2015-0043
|View full text |Cite
|
Sign up to set email alerts
|

The use of vector bootstrapping to improve variable selection precision in Lasso models

Abstract: The Lasso is a shrinkage regression method that is widely used for variable selection in statistical genetics. Commonly, K-fold cross-validation is used to fit a Lasso model. This is sometimes followed by using bootstrap confidence intervals to improve precision in the resulting variable selections. Nesting cross-validation within bootstrapping could provide further improvements in precision, but this has not been investigated systematically. We performed simulation studies of Lasso variable selection precisio… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
16
0

Year Published

2018
2018
2022
2022

Publication Types

Select...
9

Relationship

0
9

Authors

Journals

citations
Cited by 28 publications
(16 citation statements)
references
References 54 publications
0
16
0
Order By: Relevance
“…This was repeated 1000 times. The feature importance was computed as the z-score of the absolute value of the coefficient across resampling, as reported in ( Laurin et al., 2016 ).…”
Section: Methodsmentioning
confidence: 99%
“…This was repeated 1000 times. The feature importance was computed as the z-score of the absolute value of the coefficient across resampling, as reported in ( Laurin et al., 2016 ).…”
Section: Methodsmentioning
confidence: 99%
“…Predictor variable selection was performed using bootstrapped elastic net regularization [ 32 ]. Elastic net regularization is a penalized regression method, combining least absolute shrinkage and selection operator (LASSO) and ridge regression.…”
Section: Methodsmentioning
confidence: 99%
“…This is an extension of the relaxed lasso (Meinshausen, 2007). Although the elastic net does not estimate p -values to test the significance of each predictor, using cross-validation to find regression coefficients with non-zero parameter estimates can be thought of as a form of identifying which predictors are “important” (e.g., Laurin et al, 2016). Using repeated cross-validation to choose a final model attempts to derive results with an eye on generalizing to alternate samples.…”
Section: Methodsmentioning
confidence: 99%