2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) 2015
DOI: 10.1109/cvpr.2015.7299155
|View full text |Cite
|
Sign up to set email alerts
|

Understanding deep image representations by inverting them

Abstract: Image representations, from SIFT and Bag of Visual Words to Convolutional Neural Networks (CNNs), are a crucial component of almost any image understanding system. Nevertheless, our understanding of them remains limited. In this paper we conduct a direct analysis of the visual information contained in representations by asking the following question: given an encoding of an image, to which extent is it possible to reconstruct the image itself? To answer this question we contribute a general framework to invert… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

8
1,264
2

Year Published

2017
2017
2023
2023

Publication Types

Select...
7
3

Relationship

0
10

Authors

Journals

citations
Cited by 1,574 publications
(1,274 citation statements)
references
References 25 publications
8
1,264
2
Order By: Relevance
“…Analysis of the DNN-based background subtraction is needed for discussing the characteristics and the issues. Visualization methods for analyzing DNNs are proposed [28][29][30]. The authors visualized features contributing to classification by DNNs.…”
Section: Related Workmentioning
confidence: 99%
“…Analysis of the DNN-based background subtraction is needed for discussing the characteristics and the issues. Visualization methods for analyzing DNNs are proposed [28][29][30]. The authors visualized features contributing to classification by DNNs.…”
Section: Related Workmentioning
confidence: 99%
“…The top layers of the network end up-if sufficiently deep-capturing the content of the image, i.e., forming archetypal representations of the objects on which they have been trained: faces, animals, buildings, etc. (Mahendran and Vedaldi, 2014). In a symmetric movement, the potential of deep architectures to encode complex structures such as images or sound can also be used to produce new expressions of these objects: a property that has initiated a new wave of creative applications in the fine arts.…”
Section: From Shallow To Deep Neural Networkmentioning
confidence: 99%
“…In an attempt to better understand the properties of a CNN, some recent vision works have focused on analyzing their internal representations (Szegedy et al 2014;Yosinski et al 2014;Lenc and Vedaldi 2015;Mahendran and Vedaldi 2015;Zeiler and Fergus 2014;Simonyan et al 2014;Agrawal et al 2014;Zhou et al 2015;Eigen et al 2013). Some of these investigated properties of the network, like stability (Szegedy et al 2014), feature transferability (Yosinski et al 2014), equivariance, invariance and equivalence (Lenc and Vedaldi 2015), the ability to reconstruct the input (Mahendran and Vedaldi 2015) and how the number of layers, filters and parameters affects the network performance (Agrawal et al 2014;Eigen et al 2013). Zeiler and Fergus (2014) use deconvolutional networks to visualize locally optimal visual inputs for individual filters.…”
Section: Related Workmentioning
confidence: 99%