We show that propensity score matching (PSM), an enormously popular method of preprocessing data for causal inference, often accomplishes the opposite of its intended goal—thus increasing imbalance, inefficiency, model dependence, and bias. The weakness of PSM comes from its attempts to approximate a completely randomized experiment, rather than, as with other matching methods, a more efficient fully blocked randomized experiment. PSM is thus uniquely blind to the often large portion of imbalance that can be eliminated by approximating full blocking with other matching methods. Moreover, in data balanced enough to approximate complete randomization, either to begin with or after pruning some observations, PSM approximates random matching which, we show, increases imbalance even relative to the original data. Although these results suggest researchers replace PSM with one of the other available matching methods, propensity scores have other productive uses.
Recent advances in research tools for the systematic analysis of textual data are enabling exciting new research throughout the social sciences. For comparative politics, scholars who are often interested in non-English and possibly multilingual textual datasets, these advances may be difficult to access. This article discusses practical issues that arise in the processing, management, translation, and analysis of textual data with a particular focus on how procedures differ across languages. These procedures are combined in two applied examples of automated text analysis using the recently introduced Structural Topic Model. We also show how the model can be used to analyze data that have been translated into a single language via machine translation tools. All the methods we describe here are implemented in open-source software packages available from the authors.
In this study we resolve part of the confusion over how foreign aid affects armed conflict. We argue that aid shocks-severe decreases in aid revenues-inadvertently shift the domestic balance of power and potentially induce violence. During aid shocks, potential rebels gain bargaining strength vis-à-vis the government. To appease the rebels, the government must promise future resource transfers, but the government has no incentive to continue its promised transfers if the aid shock proves to be temporary. With the government unable to credibly commit to future resource transfers, violence breaks out. Using AidData's comprehensive dataset of bilateral and multilateral aid from 1981 to 2005, we evaluate the effects of foreign aid on violent armed conflict. In addition to rare-event logit analysis, we employ matching methods to account for the possibility that aid donors anticipate conflict. The results show that negative aid shocks significantly increase the probability of armed conflict onset.
This article provides theoretical and empirical solutions to two connected puzzles in the study of foreign aid and human rights: Do foreign aid donors use aid sanctions to punish repressive states, and if so, why? I show that donors impose aid sanctions selectively. Aid sanctions typically occur when repressive states do not have close political ties to aid donors, when violations have negative consequences for donors and when violations are widely publicized. Using a data set of bilateral foreign aid to 118 developing countries between 1981 and 2004, I find that variation in these factors largely accounts for the differing aid sanctions that result from objectively similar rights violations by the governments of developing countries.
We propose a simplified approach to matching for causal inference that simultaneously optimizes balance (similarity between the treated and control groups) and matched sample size. Existing approaches either fix the matched sample size and maximize balance or fix balance and maximize sample size, leaving analysts to settle for suboptimal solutions or attempt manual optimization by iteratively tweaking their matching method and rechecking balance. To jointly maximize balance and sample size, we introduce the matching frontier, the set of matching solutions with maximum possible balance for each sample size. Rather than iterating, researchers can choose matching solutions from the frontier for analysis in one step. We derive fast algorithms that calculate the matching frontier for several commonly used balance metrics. We demonstrate this approach with analyses of the effect of sex on judging and job training programs that show how the methods we introduce can extract new knowledge from existing data sets. comments on an early version were especially helpful. Software to implement the methods proposed in this article can be found at http://projects.iq.harvard.edu/frontier.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
customersupport@researchsolutions.com
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.
Copyright © 2024 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.