Human pose estimation is a key step in understanding human behavior in images and videos. Bottom-up human pose estimation methods are difficult to predict the correct pose of a person in large scenes due to the challenge of scale variation. In this paper we propose a two-stage hierarchical network that first acquires images in large scenes, and sends tracking command signals to a two-degree-of-freedom shooting platform equipped with an image sensor to track a moving target based on a motion target detection frame, and locally constrains the captured image stream according to a top-down target detection algorithm to retain only the content related to the motion target in the image. The processed images are fed into the generalized human pose estimation model for pose detection. We deployed the algorithm on a two-degree-of-freedom filming platform equipped with camera equipment and deployed the experimental platform to sport scenes to conduct detection experiments on sport figures in running and ski jumping sport scenes, using the sport figure and its nearby area as the ROI region to generate pictures or videos with the skeleton pose of the sport target to guide the sport training of the target figure. This investigation can solve the challenge of scale variation to some extent in bottom-up multi-human pose estimation, especially for large scenes where the person key points can be located more accurately. The experiments show that this investigation can meet the practical use requirements of speed and accuracy of sport figure pose detection in large scenes of daily sports.
Novel view synthesis is a long-standing problem. Despite the rapid development of neural radiance field (NeRF), in terms of rendering dynamic human body, NeRF still cannot achieve a good trade-off in precision and efficiency. In this paper, we aim at synthesizing a free-viewpoint video of an arbitrary human performers in an efficient way, only requiring a sparse number of camera views as inputs and skirting per-case fine-tuning. Recently, several works have addressed this problem by learning person-specific neural radiance fields to capture the appearance of a particular human. In parallel, some work proposed to use pixel-aligned features to generalize radiance fields to arbitrary new scenes and objects. Adopting these generalization aapprochs to human achieve reasonable rendering result. However, due to the difficulties of modeling the complex appearance of human and the dynamic sense, it is challenging to train nerf well in an efficient way. We find that the slow convergence of the human body reconstruction model is largely due to the nerf representation. In this work, we introduce a voxel grid based representation for human view synthesis, termed Voxel Grid Performer(VGP). Specifically, a sparse voxel grid is designed to represent the density and color in every sapce voxel, which enable better performance and less computation than conventional nerf optimization. We perform extensive experiments on both seen human performer and unseen human performer, demonstrating that our approach surpasses nerf-based methods on a wide variety of metrics. Code and data will be made available at https://github.com/fanzhongyi/vgp.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
customersupport@researchsolutions.com
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.
Copyright © 2025 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.