2018
DOI: 10.1007/978-3-030-01237-3_7
|View full text |Cite
|
Sign up to set email alerts
|

Modality Distillation with Multiple Stream Networks for Action Recognition

Abstract: Diverse input data modalities can provide complementary cues for several tasks, usually leading to more robust algorithms and better performance. However, while a (training) dataset could be accurately designed to include a variety of sensory inputs, it is often the case that not all modalities could be available in real life (testing) scenarios, where a model has to be deployed. This raises the challenge of how to learn robust representations leveraging multimodal data in the training stage, while considering… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

2
106
0

Year Published

2019
2019
2021
2021

Publication Types

Select...
4
2
1

Relationship

1
6

Authors

Journals

citations
Cited by 150 publications
(108 citation statements)
references
References 32 publications
2
106
0
Order By: Relevance
“…Inspired by the generalized distillation paradigm, we follow a staged learning procedure, where the "teacher" net is trained first (Step 1) and separately from the "student" (Step 2). This is in contrast with [10], where everything is learned end-to-end, but in line with [11], where separated learning steps proved to be more effective.…”
Section: Training Proceduresmentioning
confidence: 63%
See 3 more Smart Citations
“…Inspired by the generalized distillation paradigm, we follow a staged learning procedure, where the "teacher" net is trained first (Step 1) and separately from the "student" (Step 2). This is in contrast with [10], where everything is learned end-to-end, but in line with [11], where separated learning steps proved to be more effective.…”
Section: Training Proceduresmentioning
confidence: 63%
“…In this context, the closest works to our approach are [10] by Hoffman et al and [11] by Garcia et al…”
Section: Generalized Distillationmentioning
confidence: 99%
See 2 more Smart Citations
“…The hallucination network in [18] was trained on an existing modality to regress the missing modality using L 2 loss, and leveraged multiple such losses for multiple tasks. This work has been extended by Garcia et al by adding L 2 losses for reconstructing all layers of the depth network and a cross entropy distillation loss for a missing network [12]. Finally, Luo et al [26] learned the direction of distillation between modalities, considering a cosine distillation loss and a representation loss.…”
Section: Knowledge Distillationmentioning
confidence: 99%