Human skin detection, i.e. the process of discriminating "skin" and "non-skin" pixel in an image or a video, is a very important task for several applications including face detection, video surveillance, body tracking, hand gesture recognition, and many other. Skin detection has been widely studied from the research community resulting in several methods based on hand-crafted rules or deep learning. In this work we propose a novel post-processing approach for skin detectors based on trained morphological operators. The first step, consisting in skin segmentation, is performed according to an existing skin detection approach, then a second step is carried out consisting in the application of a set of morphological operators to refine the resulting mask. Extensive experimental evaluation, performed considering two different detection approaches (one based on deep learning and a handcrafted one), carried on 10 different datasets confirms the quality of the proposed method. To encourage future comparisons the MATLAB source code is freely available in the GitHub repository: https://github.com/LorisNanni.
A guiding robot aims to effectively bring people to and from specific places within environments that are possibly unknown to them. During this operation the robot should be able to detect and track the accompanied person, trying never to lose sight of her/him. A solution to minimize this event is to use an omnidirectional camera: its 360 • Field of View (FoV) guarantees that any framed object cannot leave the FoV if not occluded or very far from the sensor. However, the acquired panoramic videos introduce new challenges in perception tasks such as people detection and tracking, including the large size of the images to be processed, the distortion effects introduced by the cylindrical projection and the periodic nature of panoramic images. In this paper, we propose a set of targeted methods that allow to effectively adapt to panoramic videos a standard people detection and tracking pipeline originally designed for perspective cameras. Our methods have been implemented and tested inside a deep learning-based people detection and tracking framework with a commercial 360 • camera. Experiments performed on datasets specifically acquired for guiding robot applications and on a real service robot show the effectiveness of the proposed approach over other state-of-the-art systems. We release with this paper the acquired and annotated datasets and the open-source implementation of our method.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.