2020
DOI: 10.1109/tip.2019.2940684
|View full text |Cite
|
Sign up to set email alerts
|

Few-Shot Deep Adversarial Learning for Video-Based Person Re-Identification

Abstract: Recent years have witnessed a great development of deep learning based video person re-identification (Re-ID).A key factor for video person Re-ID is how to effectively construct discriminative video feature representations for the robustness to many complicated situations like occlusions. Recent part-based approaches employ spatial and temporal attention to extract the representative local features. While the correlations between the parts are ignored in the previous methods, to leverage the relations of diffe… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
32
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
6
3

Relationship

0
9

Authors

Journals

citations
Cited by 90 publications
(32 citation statements)
references
References 101 publications
(196 reference statements)
0
32
0
Order By: Relevance
“…In Table 1 , the results obtained with the proposed method are compared with current key works of the state-of-the-art. In particular, the comparison was performed with deep network based methods, namely, RNN [ 42 ], CNN + XQDA + MQ [ 40 ], Spatial and Temporal RNN (SPRNN) [ 43 ], Attentive Spatial-Temporal Pooling Network (ASTPN) [ 44 ], Deep Siamese Network (DSAN) [ 50 ], PersonVLAD + XQDA [ 45 ], VRNN + KissME [ 47 ], and Superpixel-Based Temporally Aligned Representation (STAR) [ 49 ]. Regarding the proposed method, 5 different versions were used for comparison.…”
Section: Resultsmentioning
confidence: 99%
See 1 more Smart Citation
“…In Table 1 , the results obtained with the proposed method are compared with current key works of the state-of-the-art. In particular, the comparison was performed with deep network based methods, namely, RNN [ 42 ], CNN + XQDA + MQ [ 40 ], Spatial and Temporal RNN (SPRNN) [ 43 ], Attentive Spatial-Temporal Pooling Network (ASTPN) [ 44 ], Deep Siamese Network (DSAN) [ 50 ], PersonVLAD + XQDA [ 45 ], VRNN + KissME [ 47 ], and Superpixel-Based Temporally Aligned Representation (STAR) [ 49 ]. Regarding the proposed method, 5 different versions were used for comparison.…”
Section: Resultsmentioning
confidence: 99%
“…In particular, the method attends human body part appearance and motion simultaneously and aggregates the extracted features via the vector of locally aggregated descriptors (VLAD) [ 46 ] aggregator. By considering the adversarial learning approach, in Reference [ 47 ] the authors presented a deep few-shot adversarial learning to produce effective video representations for video-based person re-identification, using few labelled training paired videos. In detail, the method is based on Variational Recurrent Neural Networks (VRNNs) [ 48 ], which can capture temporal dynamics by mapping video sequences into latent variables.…”
Section: Related Workmentioning
confidence: 99%
“…Recently, Deep learning [86,31,74,87,88,77,75,76,60,22,57,85,84] technique as a kind of exceedingly powerful and efficient tool is applied for the tasks of person Re-Id. Combining hand-crafted histogram features and CNN features, Wu et al [89] presented a novel feature extraction model called Feature Fusion Net (FFN) to generate a novel deep feature representation which is more discriminative and compact.…”
Section: Person Re-id Via Supervised Learningmentioning
confidence: 99%
“…As for another person REID task of video based person REID [22], [23], Wu et al [24] presented the global deep video representation learning to view-based person re-identification that aggregates local 3D features across the entire video extent. The work [25] proposed a novel few-shot deep learning approahch to video-based person REID, to learn comparable representations that are discriminative and view-invariant. The research [26] designed a visual-appearance-level and spatial-temporal-level dictionary learning approach for video-based person REID task.…”
Section: Related Work a Image Based Person Re-identificationmentioning
confidence: 99%