2019
DOI: 10.1007/978-3-030-11018-5_37
|View full text |Cite
|
Sign up to set email alerts
|

Adversarial Network Compression

Abstract: Neural network compression has recently received much attention due to the computational requirements of modern deep models. In this work, our objective is to transfer knowledge from a deep and accurate model to a smaller one. Our contributions are threefold: (i) we propose an adversarial network compression approach to train the small student network to mimic the large teacher, without the need for labels during training; (ii) we introduce a regularization scheme to prevent a trivially-strong discriminator wi… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
39
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
4
3
1

Relationship

1
7

Authors

Journals

citations
Cited by 45 publications
(39 citation statements)
references
References 57 publications
0
39
0
Order By: Relevance
“…Many papers extended this approach in different directions, such as disentangling semantic concepts [37], network compression [38] [39] [40], feature augmentation [41], image to image translation [42], and explored different losses [43] and other tricks to improve performance and stability [44] [45]. Our work relates to this body of work, as the hallucination network of our model tries to generate features from the missing modality feature space through adversarial learning.…”
Section: Adversarial Learningmentioning
confidence: 99%
See 1 more Smart Citation
“…Many papers extended this approach in different directions, such as disentangling semantic concepts [37], network compression [38] [39] [40], feature augmentation [41], image to image translation [42], and explored different losses [43] and other tricks to improve performance and stability [44] [45]. Our work relates to this body of work, as the hallucination network of our model tries to generate features from the missing modality feature space through adversarial learning.…”
Section: Adversarial Learningmentioning
confidence: 99%
“…For that we need a student-teacher adversarial framework. This has an interesting parallel in adversarial network compression [38], where the performance of a fully supervised small network can be boosted by adversarial training against a high-capacity (and better performing) teacher net. In [38], it is also observed that the student can surpass the teacher in some occasions.…”
Section: Standard Supervised Learning Has Limitations In Extracting Imentioning
confidence: 99%
“…Knowledge Distillation: Knowledge distillation (Ba and Caruana 2014) is used to transfer knowledge from teacher network to student network by the output before the softmax function (logits) or after it (soft targets), which has been popularized by (Hinton, Vinyals, and Dean 2015). As it is hard for student network with small capacity to mimic the outputs of teacher network, several researches (Belagiannis, Farshad, and Galasso 2018;Xu, Hsu, and Huang 2018) focused on using adversarial networks to replace the manually designed metric such as L1/L2 loss or KL divergence.…”
Section: Related Workmentioning
confidence: 99%
“…In recent years, many researchers resort to process-oriented methods, and many kinds of knowledge representation algorithms have been proposed (Zagoruyko and Komodakis 2016;Yim et al 2017). Empirically, the loss learned by adversarial training usually has advantages over the predetermined one in the student-teacher strategy, (Belagiannis, Farshad, and Galasso 2018) and (Xu, Hsu, and Huang 2018) proposed the GAN-based distillation approaches by introducing the discriminator to match the output distribution between teacher and student.…”
Section: Introductionmentioning
confidence: 99%
“…GANs are used in a lot of diverse applications in which generative models are involved. These include learning of data representations [19], semantic segmentation [20], teacher-student network compression [21], defending adversarial examples [22], [23], [24], and reinforcement learning [25]. The generation of training and validation material for autonomous driving systems is another use case of generative models.…”
Section: A Generative Adversarial Networkmentioning
confidence: 99%