2019
DOI: 10.48550/arxiv.1901.10912
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

A Meta-Transfer Objective for Learning to Disentangle Causal Mechanisms

Abstract: We propose to meta-learn causal structures based on how fast a learner adapts to new distributions arising from sparse distributional changes, e.g. due to interventions, actions of agents and other sources of non-stationarities. We show that under this assumption, the correct causal structural choices lead to faster adaptation to modified distributions because the changes are concentrated in one or just a few mechanisms when the learned knowledge is modularized appropriately. This leads to sparse expected grad… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
97
0
1

Year Published

2021
2021
2024
2024

Publication Types

Select...
6
3

Relationship

1
8

Authors

Journals

citations
Cited by 54 publications
(98 citation statements)
references
References 14 publications
0
97
0
1
Order By: Relevance
“…hypothesis, i.e., training and testing graphs are independently sampled from the identical distribution (Liao et al, 2020). However, in reality, such a hypothesis is hardly satisfied due to the uncontrollable generation mechanism of real data, such as data selection biases, confounder factors or other peculiarities (Bengio et al, 2019;Engstrom et al, 2019;Su et al, 2019;Hendrycks & Dietterich, 2019). The testing distribution may incur uncontrolled and unknown shifts from the training distribution, called Out-Of-Distribution (OOD) shifts (Sun et al, 2019;Krueger et al, 2021), which makes most GNN models fail to make stable predictions.…”
Section: Correlation (Star) Motifmentioning
confidence: 99%
“…hypothesis, i.e., training and testing graphs are independently sampled from the identical distribution (Liao et al, 2020). However, in reality, such a hypothesis is hardly satisfied due to the uncontrollable generation mechanism of real data, such as data selection biases, confounder factors or other peculiarities (Bengio et al, 2019;Engstrom et al, 2019;Su et al, 2019;Hendrycks & Dietterich, 2019). The testing distribution may incur uncontrolled and unknown shifts from the training distribution, called Out-Of-Distribution (OOD) shifts (Sun et al, 2019;Krueger et al, 2021), which makes most GNN models fail to make stable predictions.…”
Section: Correlation (Star) Motifmentioning
confidence: 99%
“…As a result, there has been a surge in interest in differentiable structure learning and the combination of deep learning and causal inference . Such methods define a structural causal model with smoothly differentiable parameters that are adjusted to fit observational data (Zheng et al, 2018;Yu et al, 2019;Zheng et al, 2020;Bengio et al, 2019;Lorch et al, 2021;Annadani et al, 2021), although some methods can accept interventional data, thereby significantly improving the identification of the underlying data-generating process Lippe et al, 2021). However, the improvement critically depends on the experiments and interventions available.…”
Section: Introductionmentioning
confidence: 99%
“…Uncovering and understanding causal mechanisms is an important problem not only in machine learning [3,31,39] but also in various scientific disciplines such as computational biology [10,28,37], epidemiology [36,46], and economics [15,31]. A common task of interest is causal structure learning [31,33] which aims at learning a directed acyclic graph (DAG) in which edges represent causal relations between variables.…”
Section: Introductionmentioning
confidence: 99%
“…Finding the right DAG is challenging as the solution space grows super-exponentially with the number of variables. A promising new direction are continuous-optimization methods [3,4,18,51,52,53,54] that are more computationally efficient than previous score-based and constraint-based methods [12,33] while leveraging the expressiveness of neural networks as function approximators. To restrict the search space to acyclic graphs, Zheng et al [52] first proposed to view the search as a constrained optimization problem using an augmented Lagrangian procedure to solve it.…”
Section: Introductionmentioning
confidence: 99%