2021
DOI: 10.48550/arxiv.2104.13369
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Explaining in Style: Training a GAN to explain a classifier in StyleSpace

Abstract: Image classification models can depend on multiple different semantic attributes of the image. An explanation of the decision of the classifier needs to both discover and visualize these properties. Here we present StylEx, a method for doing this, by training a generative model to specifically explain multiple attributes that underlie classifier decisions. A natural source for such attributes is the StyleSpace of StyleGAN, which is known to generate semantically meaningful dimensions in the image. However, bec… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
15
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
6
1
1

Relationship

1
7

Authors

Journals

citations
Cited by 14 publications
(15 citation statements)
references
References 34 publications
0
15
0
Order By: Relevance
“…Indeed, for some attributes, even finding a disentangled latent direction is infeasible. Furthermore, similar to other methods which rely on StyleGAN [7,29], our method obtains better results when operating on images within the domain used to train the GAN. This limitation stems in part from to the inability of current GAN Inversion methods to reconstruct out-of-domain images while preserving latent semantics.…”
Section: Discussionmentioning
confidence: 52%
See 1 more Smart Citation
“…Indeed, for some attributes, even finding a disentangled latent direction is infeasible. Furthermore, similar to other methods which rely on StyleGAN [7,29], our method obtains better results when operating on images within the domain used to train the GAN. This limitation stems in part from to the inability of current GAN Inversion methods to reconstruct out-of-domain images while preserving latent semantics.…”
Section: Discussionmentioning
confidence: 52%
“…In the context of discriminative tasks, several recent methods have proposed to utilize GANs for additional purposes. Lang et al [29] used StyleGAN [27] to visualize counterfactual examples for explaining a pretrained classifier's predictions. Chai et al [7] used style-mixing in the fine-layers of StyleGAN to generate augmentations that are ensembled together at test-time.…”
Section: Related Workmentioning
confidence: 99%
“…In the first step, we jointly train a generator G and an encoder E. We use the popular Style-GAN2 generator [16], and train it to produce realistic images using the original objective as have been employed by Karras et al [12,15,16], denoted L GAN . The encoder is trained to reconstruct both real and synthesized images, similar to the work of Lang et al [19]. Let G n (z) be an unconditionally generated image from normally distributed noise vector z, we denote its reconstruction as G(E(G n (z))), where G(w) refers to applying the generator over latent code w (i.e.…”
Section: Generative-based Self-filteringmentioning
confidence: 99%
“…GANalyze [8] takes advantages of the GAN-based model to visualize what a CNN model learns about high-level cognitive properties. StylEx [19] proposes to incorporate the classifier into the training process of StyleGAN and learn a classifier-specific StyleSpace. Sauer and Geiger [26] propose to disentangle object shape, object texture and background in the image generation process and generate structured conterfacturals which help improve the robustness and interpretability of classifiers.…”
Section: Generative Counterfactual Imagesmentioning
confidence: 99%
“…On one hand, it is difficult to sample valid counterfactuals from high-dimensional image manifold by simply altering pixels. To address this difficulty, some recent studies consider to exploit image generation techniques to produce counterfactuals [7,19,30]. On the other hand, even though image counterfactuals can be sampled using generators, it is sometimes still difficult to explain what attributes or concepts are altered.…”
Section: Introductionmentioning
confidence: 99%