2019
DOI: 10.48550/arxiv.1911.10500
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Causality for Machine Learning

Abstract: Graphical causal inference as pioneered by Judea Pearl arose from research on artificial intelligence (AI), and for a long time had little connection to the field of machine learning. This article discusses where links have been and should be established, introducing key concepts along the way. It argues that the hard open problems of machine learning and AI are intrinsically related to causality, and explains how the field is beginning to understand them.

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

2
130
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
3
2
2

Relationship

0
7

Authors

Journals

citations
Cited by 87 publications
(132 citation statements)
references
References 54 publications
2
130
0
Order By: Relevance
“…We show that ML classifiers (Logistic regression and LSTM), when used by themselves directly on time-series measurements are dumb to the temporal/ causal-structure in the data. This fact has also been discussed in existing literature [1,2]. When time-series values were directly passed to the classifiers, LR and LSTM, they failed to learn any causal-structure characteristics from the data.…”
Section: Discussion Concluding Remarks and Future Research Directionsmentioning
confidence: 54%
See 2 more Smart Citations
“…We show that ML classifiers (Logistic regression and LSTM), when used by themselves directly on time-series measurements are dumb to the temporal/ causal-structure in the data. This fact has also been discussed in existing literature [1,2]. When time-series values were directly passed to the classifiers, LR and LSTM, they failed to learn any causal-structure characteristics from the data.…”
Section: Discussion Concluding Remarks and Future Research Directionsmentioning
confidence: 54%
“…What if I had acted in a different way? Machine intelligence is still far away from answering these kind of questions [1,2].…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…Most prior approaches assume that inputs are structured as disentangled variables [6,16,26,36,48,47], which often does not hold in domains with high-dimensional inputs, i.e., images. While Lopez-Paz et al [29] demonstrated the possibility of observational causal discovery from high-dimensional images, combining causal models and representation learning in such domains still remains an open problem [46]. Hence, we instead explore the approach of regularizing a policy that operates on high-dimensional states.…”
Section: Vq-vae Encodermentioning
confidence: 99%
“…In order to address this causal confusion problem, one can consider causal discovery approaches to deduce the cause-effect relationships from observational data [26,48]. However, it is difficult to apply these approaches to domains with high-dimensional inputs, as (i) causal discovery from observational data is impossible in general without certain conditions 3 [38], and (ii) these domains usually do not satisfy the assumption that inputs are structured into random variables connected by a causal graph, e.g., objects in images [29,46]. To address these limitations, de Haan et al [12] recently proposed a method that learns a policy on top of disentangled representations from a β-VAE encoder [19] with random masking, and infers an optimal causal mask during the environment interaction by querying interactive experts [43] or environment returns.…”
Section: Introductionmentioning
confidence: 99%