2003
DOI: 10.1177/0002716203254879
|View full text |Cite
|
Sign up to set email alerts
|

Nonexperimental Versus Experimental Estimates of Earnings Impacts

Abstract: To assess nonexperimental (NX) evaluation methods in the context of welfare, job training, and employment services programs, the authors reexamined the results of twelve case studies intended to replicate impact estimates from an experimental evaluation by using NX methods. They found that the NX methods sometimes came close to replicating experimentally derived results but often produced estimates that differed by policy-relevant margins, which the authors interpret as estimates of bias. Although the authors … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

3
289
0
3

Year Published

2008
2008
2016
2016

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 286 publications
(295 citation statements)
references
References 16 publications
3
289
0
3
Order By: Relevance
“…However, some general guidance can be obtained. Glazerman et al (2003) find that one-to-one propensity score matching performs better than other propensity score matching methods or non-propensity score matching and that standard econometric selection correction procedures, such as instrumental variables or the Heckman selection correction, tend to perform poorly. As discussed earlier, their results also show that combining methods, such as matching and covariance adjustment, is better than using those methods individually.…”
Section: Evaluation Of Matching Methodsmentioning
confidence: 92%
See 2 more Smart Citations
“…However, some general guidance can be obtained. Glazerman et al (2003) find that one-to-one propensity score matching performs better than other propensity score matching methods or non-propensity score matching and that standard econometric selection correction procedures, such as instrumental variables or the Heckman selection correction, tend to perform poorly. As discussed earlier, their results also show that combining methods, such as matching and covariance adjustment, is better than using those methods individually.…”
Section: Evaluation Of Matching Methodsmentioning
confidence: 92%
“…The matching method reduces large covariate bias between the treated and control groups, and the regression is used to adjust for any small residual biases and to increase efficiency. These "bias-corrected" matching methods have been found by Abadie and Imbens (2006) and Glazerman, Levy, and Myers (2003) to work well in practice, using simulated and actual data. Rubin (1973bRubin ( , 1979, Rubin and Thomas (2000), and Ho et al (2007) show that models based on matched data are much less sensitive to model misspecification and more robust than are models fit in the full data sets.…”
Section: Analysis Of Outcome Data After Matchingmentioning
confidence: 94%
See 1 more Smart Citation
“…When comparing average differences across multiple randomised and non-randomised studies using meta-analysis, the evidence is mixed. One prominent article found significant differences in estimated effects between 12 replicated randomised and non-randomised studies (Glazerman et al 2003), whereas others have suggested that the differences in results are almost zero overall, whether or not the individual study differences are themselves zero Wilson 1993, Heinsman andShadish 1996). Hansen et al (2011) note, based on an assessment of within-study comparisons, that appropriate knowledge of the participation decision process is key to estimating unbiased effects.…”
Section: Risk Of Bias Assessment In Quasi-experimental Designsmentioning
confidence: 99%
“…This includes use of propensity scores in weighted regression and in earlier techniques, such as (i) creating match sets or (ii) computing weighted contrasts between treatment and control groups. See Other references of interest include Arceneaux, Gerber, and Green (2006), Glazerman, Levy, and Myers (2003), Peikes, Moreno, and Orzol (2008), Wilde and Hollister (2007). These authors point to serious weaknesses in the propensity-score methods that have been used for program evaluation.…”
Section: Literature Reviewmentioning
confidence: 99%