2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2018
DOI: 10.1109/icassp.2018.8461758
|View full text |Cite
|
Sign up to set email alerts
|

Incorporating Scalability in Unsupervised Spatio- Temporal Feature Learning

Abstract: Deep neural networks are efficient learning machines which leverage upon a large amount of manually labeled data for learning discriminative features. However, acquiring substantial amount of supervised data, especially for videos can be a tedious job across various computer vision tasks. This necessitates learning of visual features from videos in an unsupervised setting. In this paper, we propose a computationally simple, yet effective, framework to learn spatio-temporal feature embedding from unlabeled vide… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
4
0

Year Published

2020
2020
2020
2020

Publication Types

Select...
1
1

Relationship

1
1

Authors

Journals

citations
Cited by 2 publications
(4 citation statements)
references
References 26 publications
0
4
0
Order By: Relevance
“…We investigate the contributions of the temporal losses to the re-identification TCPL (L) EUG [30] One-Shot Prog. [29] (b) Fig. 4.…”
Section: Methodsmentioning
confidence: 99%
See 3 more Smart Citations
“…We investigate the contributions of the temporal losses to the re-identification TCPL (L) EUG [30] One-Shot Prog. [29] (b) Fig. 4.…”
Section: Methodsmentioning
confidence: 99%
“…(1) using the unlabeled data efficiently is important, ( TCPL EUG [30] One-Shot Progressive [29] (b) performance. In order to do that, we performed experiments with different values of λ (higher value indicates larger weight on the temporal losses) and present the results on the DukeMTMC-VideoReID dataset in Fig.…”
Section: Methodsmentioning
confidence: 99%
See 2 more Smart Citations