2021
DOI: 10.1214/21-ejs1825
|View full text |Cite
|
Sign up to set email alerts
|

Multicarving for high-dimensional post-selection inference

Abstract: We consider post-selection inference for high-dimensional (generalized) linear models. Data carving from Fithian, Sun and Taylor [10] is a promising technique to perform this task. However, it suffers from the instability of the model selector and hence, may lead to poor replicability, especially in high-dimensional settings. We propose the multicarve method inspired by multisplitting to improve upon stability and replicability. Furthermore, we extend existing concepts to group inference and illustrate the app… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
10
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
4

Relationship

0
9

Authors

Journals

citations
Cited by 11 publications
(10 citation statements)
references
References 32 publications
0
10
0
Order By: Relevance
“…Third, p-values computed under our asymptotic selective likelihood can be easily adapted to address the p-value lottery problem. In this sense, our procedure is related to the multi-carving approach for improved replication [35], but yields us a much faster, sampling-free solution to the same problem.…”
Section: Contributions and Other Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Third, p-values computed under our asymptotic selective likelihood can be easily adapted to address the p-value lottery problem. In this sense, our procedure is related to the multi-carving approach for improved replication [35], but yields us a much faster, sampling-free solution to the same problem.…”
Section: Contributions and Other Related Workmentioning
confidence: 99%
“…Finally, our procedure can be easily adapted to address the "p-value lottery" problem, which arises frequently with model selection on random splits of data. The proposal serves as an efficient alternative to multi-splitting in [11], and multi-carving in [35]. Multi-carving is a more powerful version of multi-splitting, and is based on techniques that are known as carving [13,26].…”
Section: Introductionmentioning
confidence: 99%
“…To develop the test, we leverage the selective inference framework, which has been applied extensively in high-dimensional linear modeling (Lee et al, 2016;Tibshirani et al, 2016;Fithian et al, 2014;Rügamer et al, 2022;Schultheiss et al, 2021;Taylor & Tibshirani, 2018;Charkhi & Claeskens, 2018;Yang et al, 2016;Loftus & Taylor, 2014), changepoint detection (Jewell et al, 2022;Hyun et al, 2021Hyun et al, , 2018Chen et al, 2021b;Le Duy & Takeuchi, 2021;Duy et al, 2020;Benjamini et al, 2019), and clustering (Zhang et al, 2019;Gao et al, 2020;Watanabe & Suzuki, 2021). The key insight behind selective inference is as follows: to obtain a valid test of H 0 , we need to condition on the aspect of the data that led us to test it.…”
Section: Introductionmentioning
confidence: 99%
“…The idea of conditioning upon the observed selection, previously explored in [7,17,11], among others, discards bias from model selection through p-values, confidence intervals, credible intervals based on conditional probabilities. The estimator we propose here implements the principles of data carving [4,12,10,14] within the conditional framework. Data carving resembles data-splitting in the selection stage, because model selection operates on only a subset of the full data.…”
Section: Introductionmentioning
confidence: 99%