2017
DOI: 10.1016/j.patrec.2016.12.014
|View full text |Cite
|
Sign up to set email alerts
|

People silhouette extraction from people detection bounding boxes in images

Abstract: In many applications such as video surveillance or autonomous vehicles, people detection is a key element, often based on feature extraction and combined with supervised classification. Usually, output of these methods is in the form of a bounding-box containing an extracted people along with the background. But in specific application contexts, this bounding box information is not sufficient and a precise segmentation of people silhouette is needed inside the bounding box. For videos, this is actually solved … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
4
0

Year Published

2017
2017
2023
2023

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 13 publications
(4 citation statements)
references
References 23 publications
0
4
0
Order By: Relevance
“…The polygonization step is adaptive and works completely onthe-fly for each binary image. One of the novelties of this study is that; instead of using conventional methods such as the difference between video images and background subtraction techniques to determine the silhouette [44], [45], [46], [47], a deep learning architecture, Yolact++, is modified, and the image information obtained from the middleware is used. This step exploits the power of the architecture to represent the silhouette as a whole with greater accuracy than that of conventional methods.…”
Section: The Proposed Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…The polygonization step is adaptive and works completely onthe-fly for each binary image. One of the novelties of this study is that; instead of using conventional methods such as the difference between video images and background subtraction techniques to determine the silhouette [44], [45], [46], [47], a deep learning architecture, Yolact++, is modified, and the image information obtained from the middleware is used. This step exploits the power of the architecture to represent the silhouette as a whole with greater accuracy than that of conventional methods.…”
Section: The Proposed Methodsmentioning
confidence: 99%
“…On the other hand, motionbased feature extraction methods focus on extracting information related to the motion of objects or people in the video and these methods extract features that describe the temporal changes in the position, velocity, and acceleration of body parts in the video [5], [6], [7], [8], [9]. They are also well-suited for recognizing actions that involve significant motion, such as waving, punching, and kicking [44], [51], [52], [53], [54], [55], [56], [57].…”
Section: Related Work and Creation Of Silhouettesmentioning
confidence: 99%
“…Finally, the bounding boxes are refined by the elimination of duplicate detections and rescoring the boxes based on other objects on the scene using SVMs ( Girshick et al, 2014 ). The bounding box is a rectangular box located around the objects in order to represent their detection ( Coniglio et al, 2017 , Lézoray and Grady, 2012 ). The resulting object detection datasets are images with tags used to classify different categories ( Deng et al, 2009 , Everingham et al, 2010 ).…”
Section: Digital Image Processing For Object Detectionmentioning
confidence: 99%
“…Truong Cong et al [14,15] report a foreground estimation and person re-identification approach, but it is not clear how it can deal with stationary people, nor they consider different postures. More recently Coniglio et al [16] present an interesting approach based on shape priors, HOG and multiple SVMs but it is not clear how they deal with different poses.…”
Section: Introductionmentioning
confidence: 99%