This paper studies probability distributions of penultimate activations of classification networks. We show that, when a classification network is trained with the cross-entropy loss, its final classification layer forms a Generative-Discriminative pair with a generative classifier based on a specific distribution of penultimate activations. More importantly, the distribution is parameterized by the weights of the final fully-connected layer, and can be considered as a generative model that synthesizes the penultimate activations without feeding input data. We empirically demonstrate that this generative model enables stable knowledge distillation in the presence of domain shift, and can transfer knowledge from a classifier to variational autoencoders and generative adversarial networks for class-conditional image generation.
INTRODUCTIONDeep neural networks have achieved remarkable success in image classification [10,13]. In most of these networks, an input image is first processed by multiple layers of neurons, whose final output, called penultimate activations, is in turn fed to the last fully connected layer that conducts classification. These networks are typically trained in an end-to-end manner by minimizing the cross-entropy loss. The penultimate activations are the deepest image representation of the networks and have proven to be useful for various purposes besides classification such as image retrieval [45], semantic segmentation [26], and general image description of unseen classes [35].This paper studies the penultimate activations of classifica-* The two authors equally contributed. This work was done while Minkyo Seo was visiting Kakao as a research intern.