2018
DOI: 10.1111/insr.12253
|View full text |Cite
|
Sign up to set email alerts
|

Comparing Inference Methods for Non‐probability Samples

Abstract: Summary Social and economic scientists are tempted to use emerging data sources like big data to compile information about finite populations as an alternative for traditional survey samples. These data sources generally cover an unknown part of the population of interest. Simply assuming that analyses made on these data are applicable to larger populations is wrong. The mere volume of data provides no guarantee for valid inference. Tackling this problem with methods originally developed for probability sampli… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

2
28
0
4

Year Published

2020
2020
2022
2022

Publication Types

Select...
5
3

Relationship

0
8

Authors

Journals

citations
Cited by 48 publications
(39 citation statements)
references
References 65 publications
2
28
0
4
Order By: Relevance
“…In general, our findings support the view given in [65] that ML methods can usefully be used to remove selection bias when dealing with non-probability samples. Prior research has shown that PSA successfully removes bias in some situations but at the cost of increasing the variance of the estimates ( [10]; [11]).…”
Section: Resultssupporting
confidence: 89%
See 1 more Smart Citation
“…In general, our findings support the view given in [65] that ML methods can usefully be used to remove selection bias when dealing with non-probability samples. Prior research has shown that PSA successfully removes bias in some situations but at the cost of increasing the variance of the estimates ( [10]; [11]).…”
Section: Resultssupporting
confidence: 89%
“…In this respect, [65] reviewed existing inference methods to correct for selection bias in nonprobability samples. These authors considered a situation where only a nonprobability sample is available and compared a range of predictive inference methods (pseudo-designbased and model-based) in a general framework.…”
Section: Discussionmentioning
confidence: 99%
“…Other applications of machine learning algorithms in PSA involve their use in nonresponse adjustments; more precisely, they have been studied using Random Forests as propensity predictors [21]. Regarding the nonprobability sampling context covered in this study, [12] presented a simulation study using decision trees, k-Nearest Neighbors, Naive Bayes, Random Forests, and a Gradient Boosting Machine that support the view given in [6] about machine learning methods being used for removing selection bias in nonprobability samples. All of those algorithms, along with Discriminant Analysis and Model Averaged Neural Networks, will be used for propensity estimation in this study.…”
Section: Propensity Score Adjustmentmentioning
confidence: 74%
“…In order to correct this selection bias produced by non-random selection mechanisms, some inference procedures are proposed in the literature. A first class of methods includes statistical models aiming to predict the non-sampled units of the population [4][5][6]. Specifying an appropriate super-population model capable of learning the variation of the target variables is important for this model-based method.…”
Section: Introductionmentioning
confidence: 99%
“…A good overview of the various methods is given in Elliott and Valliant (2017). There are three important approaches: the pseudo-design based inference (or pseudo-randomisation (Buelens et al, 2018)), statistical matching and predictive inference.…”
Section: Introductionmentioning
confidence: 99%