2019 International Joint Conference on Neural Networks (IJCNN) 2019
DOI: 10.1109/ijcnn.2019.8851734
|View full text |Cite
|
Sign up to set email alerts
|

Pose estimator and tracker using temporal flow maps for limbs

Abstract: For human pose estimation in videos, it is significant how to use temporal information between frames. In this paper, we propose temporal flow maps for limbs (TML) and a multi-stride method to estimate and track human poses. The proposed temporal flow maps are unit vectors describing the limbs' movements. We constructed a network to learn both spatial information and temporal information end-to-end. Spatial information such as joint heatmaps and part affinity fields is regressed in the spatial network part, an… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
13
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
3
2
2

Relationship

0
7

Authors

Journals

citations
Cited by 36 publications
(13 citation statements)
references
References 25 publications
0
13
0
Order By: Relevance
“…In addition, the feature representation should (iii) focus more on appearance rather than location, since objects can move in time. Normally, such a feature embedding can be learned, e.g., via a side branch trained with labelled identities across frames as the supervision signal, as is evidenced in several existing works [5,19,50]. Although no such annotations are accessible here, we find that instance-level annotation on images already provides sufficient information to achieve the above goals, i.e., to distinguish which pixels belong to the same instance, and which are from different ones.…”
Section: Instance Contrastive Lossmentioning
confidence: 84%
“…In addition, the feature representation should (iii) focus more on appearance rather than location, since objects can move in time. Normally, such a feature embedding can be learned, e.g., via a side branch trained with labelled identities across frames as the supervision signal, as is evidenced in several existing works [5,19,50]. Although no such annotations are accessible here, we find that instance-level annotation on images already provides sufficient information to achieve the above goals, i.e., to distinguish which pixels belong to the same instance, and which are from different ones.…”
Section: Instance Contrastive Lossmentioning
confidence: 84%
“…The AP metric follows the same protocol of human instances, but to the best of our knowledge no previous method has evaluated AP on ApolloCar3D [58] without leveraging 3D information. Hence, we include a study PoseTrack 2018 MOTA FPS openSVAI [28] 54.5 -MIPAL [26] 54.9 -Miracle [30] 57.4 -MSRA/FlowTrack [19] 61.4 0.7 OpenPifPaf (ours) 61.7 12.2 (a)…”
Section: Resultsmentioning
confidence: 99%
“…PoseTrack 2017 MOTA FPS STAF [27] 53.8 3 MIPAL [26] 54.5 -MSRA/FlowTrack [19] 57.8 0.7 HRNet [20] 57.9 -LightTrack [29] 58.0 -OpenPifPaf (ours) 60.6 12.2 KeyTrack [32] 61.2 1.0 DetTrack (offline) [67] 64.1 -…”
Section: Resultsmentioning
confidence: 99%
See 2 more Smart Citations