2018
DOI: 10.48550/arxiv.1812.10305
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Spatial and Temporal Mutual Promotion for Video-based Person Re-identification

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

0
4
0

Year Published

2019
2019
2019
2019

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(4 citation statements)
references
References 12 publications
0
4
0
Order By: Relevance
“…Learning a discriminative clip-level feature is crucial to video-based person re-id. Most previous work dedicated to aggregating frame features vectors across temporal dimension into a clip-level feature [6,16,17,18,35]. In [35], the authors extracts and aggregates the temporal and spatial information between consecutive frames simultaneously with one-stream Neural Network.…”
Section: Related Workmentioning
confidence: 99%
See 3 more Smart Citations
“…Learning a discriminative clip-level feature is crucial to video-based person re-id. Most previous work dedicated to aggregating frame features vectors across temporal dimension into a clip-level feature [6,16,17,18,35]. In [35], the authors extracts and aggregates the temporal and spatial information between consecutive frames simultaneously with one-stream Neural Network.…”
Section: Related Workmentioning
confidence: 99%
“…Liao et al [16] employed a succession of 3D convolution kernel pre-trained on kinetics to extract spatial and temporal features simultaneously from a video volume, which keeps the intra-clip consistency and learns the context information of local appearance patch. On the other hand, temporal alignment is a key point to temporal pooling performance [15,18,24]. Li et al [15] create a compact encoding of the video that exploits useful partial information in each frame.…”
Section: Related Workmentioning
confidence: 99%
See 2 more Smart Citations