2021
DOI: 10.48550/arxiv.2110.03109
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Consistent Counterfactuals for Deep Models

Emily Black,
Zifan Wang,
Matt Fredrikson
et al.

Abstract: Counterfactual examples are one of the most commonly-cited methods for explaining the predictions of machine learning models in key areas such as finance and medical diagnosis. Counterfactuals are often discussed under the assumption that the model on which they will be used is static, but in deployment models may be periodically retrained or fine-tuned. This paper studies the consistency of model prediction on counterfactual examples in deep networks under small changes to initial training conditions, such as… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

1
5
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
3

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(6 citation statements)
references
References 28 publications
1
5
0
Order By: Relevance
“…However, this work focuses on small changes to the model, e.g., retraining on some data drawn from the same distribution, or minor changes to the hyperparameters, keeping the underlying data mostly similar. Such small changes to the model are in fact quite common in several applications and occur frequently in practice [12][13][14][15].…”
Section: Methodsmentioning
confidence: 99%
See 2 more Smart Citations
“…However, this work focuses on small changes to the model, e.g., retraining on some data drawn from the same distribution, or minor changes to the hyperparameters, keeping the underlying data mostly similar. Such small changes to the model are in fact quite common in several applications and occur frequently in practice [12][13][14][15].…”
Section: Methodsmentioning
confidence: 99%
“…In [6,10,18], the authors argue that counterfactuals that lie on the data manifold are likely to be more robust than the closest counterfactuals, but the focus is more on generating counterfactuals that specifically lie on the data manifold (which may not always be sufficient for robustness). Despite researchers arguing that robustness is an important desideratum of local explanation methods [13], the problem of generating robust counterfactuals has been less explored, with the notable exceptions of some recent works [12,14,22]. In [12,14], the authors propose algorithms that aim to find the closest counterfactuals that are also robust (with demonstration on linear models and neural networks).…”
Section: Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…Prior works have focused on determining the extent to which recourses remain robust to the choice of the underlying model (Pawelczyk et al, 2020b;Black et al, 2021), shifts or changes in the underlying models (Rawal et al, 2021;Upadhyay et al, 2021), or small perturbations to the input instances (Artelt et al, 2021;Dominguez-Olmedo et al, 2021;Slack et al, 2021).…”
Section: Robustness Of Algorithmic Recoursementioning
confidence: 99%
“…lead to favourable classification outcomes) for all plausible individuals similar to the individual seeking recourse. We refer to this notion of robustness as the adversarial robustness of recourse, in order to distinguish it from other robustness considerations previously studied in the recourse literature (e.g., robustness with respect to changes to the decision-making classifier [29,36,3]), and as a reference to the adversarial robustness literature, which considers robustness of prediction precisely against uncertainty in the features of the data.…”
Section: Introductionmentioning
confidence: 99%