2022
DOI: 10.1007/978-3-031-19842-7_38
|View full text |Cite
|
Sign up to set email alerts
|

BodySLAM: Joint Camera Localisation, Mapping, and Human Motion Tracking

Abstract: Robust, fast, and accurate human state -6D pose and posture -estimation remains a challenging problem. For real-world applications, the ability to estimate the human state in real-time is highly desirable. In this paper, we present BodySLAM++, a fast, efficient, and accurate human and camera state estimation framework relying on visualinertial data. BodySLAM++ extends an existing visual-inertial state estimation framework, OKVIS2, to solve the dual task of estimating camera and human states simultaneously. Our… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2

Citation Types

0
2
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
3
3
1

Relationship

0
7

Authors

Journals

citations
Cited by 12 publications
(2 citation statements)
references
References 57 publications
0
2
0
Order By: Relevance
“…Several current methods focus on positioning humans in a pre‐scanned 3D scene [GMSP21; HYH*22; HCTB19] and on simultaneous estimation of human poses and objects humans interact with [WY21; CHY*19; YHT*22]. Different setups from ours assume an RGB‐D sensor [ZZB*21b] or a moving camera [ZMZ*22; LYZ*21; HLL22; LBX*22] to estimating the scene geometry. Recent methods integrate physics‐based constraints into monocular 3D human motion capture and mitigate foot‐floor penetration and other severe artefacts [SGXT20; SGX*21].…”
Section: Related Workmentioning
confidence: 99%
“…Several current methods focus on positioning humans in a pre‐scanned 3D scene [GMSP21; HYH*22; HCTB19] and on simultaneous estimation of human poses and objects humans interact with [WY21; CHY*19; YHT*22]. Different setups from ours assume an RGB‐D sensor [ZZB*21b] or a moving camera [ZMZ*22; LYZ*21; HLL22; LBX*22] to estimating the scene geometry. Recent methods integrate physics‐based constraints into monocular 3D human motion capture and mitigate foot‐floor penetration and other severe artefacts [SGXT20; SGX*21].…”
Section: Related Workmentioning
confidence: 99%
“…In contrast, our work recovers human trajectories for in-the-wild videos, in which camera motion is uncontrolled, and the scene reconstruction is limited or non-existent. [15] operate on monocular sequences, but the extent of results is limited to a single unoccluded person slowly walking in an indoor studio. We demonstrate our approach on PoseTrack, a complex in-the-wild dataset, which includes videos with a large number of people in various environments.…”
Section: Related Workmentioning
confidence: 99%