2022
DOI: 10.1162/tacl_a_00511
|View full text |Cite
|
Sign up to set email alerts
|

Causal Inference in Natural Language Processing: Estimation, Prediction, Interpretation and Beyond

Abstract: A fundamental goal of scientific research is to learn about causal relationships. However, despite its critical role in the life and social sciences, causality has not had the same importance in Natural Language Processing (NLP), which has traditionally placed more emphasis on predictive tasks. This distinction is beginning to fade, with an emerging area of interdisciplinary research at the convergence of causal inference and language processing. Still, research on causality in NLP remains scattered across dom… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
36
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
2
1
1

Relationship

0
8

Authors

Journals

citations
Cited by 85 publications
(37 citation statements)
references
References 90 publications
1
36
0
Order By: Relevance
“…Grimmer, Roberts and Stewarts’ chapter on causal analysis highlights critical principles for causal identification and articulates challenges associated with the accuracy of causal modeling assumptions, like the Stable Unit Treatment Value Assumption (SUTVA), which requires that the response of a particular unit depends on the treatment to which it was assigned, not the treatments of those around it. They emphasize the importance of experiments, which they illustrate with some of their own pathbreaking work in this area (Egami et al 2018; Grimmer 2015). I argue below that modern transformer methods could contribute most to this area, by enabling the automated contruction of text manifesting precise semantic variation with generative models, and the simulation of text understanding and reception with encoding models.…”
Section: Notable Strengthsmentioning
confidence: 92%
See 2 more Smart Citations
“…Grimmer, Roberts and Stewarts’ chapter on causal analysis highlights critical principles for causal identification and articulates challenges associated with the accuracy of causal modeling assumptions, like the Stable Unit Treatment Value Assumption (SUTVA), which requires that the response of a particular unit depends on the treatment to which it was assigned, not the treatments of those around it. They emphasize the importance of experiments, which they illustrate with some of their own pathbreaking work in this area (Egami et al 2018; Grimmer 2015). I argue below that modern transformer methods could contribute most to this area, by enabling the automated contruction of text manifesting precise semantic variation with generative models, and the simulation of text understanding and reception with encoding models.…”
Section: Notable Strengthsmentioning
confidence: 92%
“…Perhaps the most unique contribution from Text as Data is in the penultimate section on inference, or drawing reasoned insights from observed text to that which is unobserved (e.g., written in the future, from other places or for other purposes). The authors, leaders in causal analysis with text (Feder et al 2021), clearly distinguish between prediction, or the anticipation of future events, and causal inference that requires conditional or counterfactual prediction. Their principles of prediction are clear and elegant (e.g., "predictive features do not have to cause the outcome", "it can be difficult to apply prediction to policy-making").…”
Section: Causal Analysis Of Textmentioning
confidence: 99%
See 1 more Smart Citation
“…In addition, the combination of causal inference and other research fields also exhibits complementary strengths, such as computer vision (Tang et al 2020;Liu et al 2022a), graph learning (Ma et al 2022;Chu, Rathbun, and Li 2021), natural language processing (Feder et al 2022;Liu et al 2022b), and so on. The involved causal analysis helps improve the model's capability of discovering and resolving the underlying system beyond the statistical relationships learned from observational data.…”
Section: Related Workmentioning
confidence: 99%
“…We demonstrate that widely recognized, emotion-laden events have immediate but no lasting effects on interpretations of immigration and that it takes prolonged periods of rupture to change them. This event analysis exemplifies how researchers can use the output of computational text models to trace causal claims within, for example, the traditional regression model framework (Egami et al 2018, Roberts et al 2020, Fong & Grimmer 2019, Feder et al 2021, Gencoglu & Gruber 2020.…”
Section: Introductionmentioning
confidence: 98%