2019
DOI: 10.3390/s19194092
|View full text |Cite
|
Sign up to set email alerts
|

Multi-View Fusion-Based 3D Object Detection for Robot Indoor Scene Perception

Abstract: To autonomously move and operate objects in cluttered indoor environments, a service robot requires the ability of 3D scene perception. Though 3D object detection can provide an object-level environmental description to fill this gap, a robot always encounters incomplete object observation, recurring detections of the same object, error in detection, or intersection between objects when conducting detection continuously in a cluttered room. To solve these problems, we propose a two-stage 3D object detection al… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
25
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
3

Relationship

0
7

Authors

Journals

citations
Cited by 30 publications
(25 citation statements)
references
References 26 publications
0
25
0
Order By: Relevance
“…Our asynchronous processing of dense 3D reconstruction and object annotation supports the usage of CNNs with long and varying inference time. Thereby, we improve the state of the art, as related work either requires hard-coding of the number of frames that can be interpreted [13,16,20], or only achieves interactive frame rates when using networks with short inference time [17]. Furthermore, our presented approach avoids degradation of the 3D reconstruction due to the asynchronous computation of 3D reconstruction and object detection, both with respect to time and resources.…”
Section: Discussionmentioning
confidence: 99%
See 2 more Smart Citations
“…Our asynchronous processing of dense 3D reconstruction and object annotation supports the usage of CNNs with long and varying inference time. Thereby, we improve the state of the art, as related work either requires hard-coding of the number of frames that can be interpreted [13,16,20], or only achieves interactive frame rates when using networks with short inference time [17]. Furthermore, our presented approach avoids degradation of the 3D reconstruction due to the asynchronous computation of 3D reconstruction and object detection, both with respect to time and resources.…”
Section: Discussionmentioning
confidence: 99%
“…With our novel filter pipeline, our framework achieves fast and efficient object classification by highly efficient, generic modification and filtering of the predicted 2D bounding boxes. This makes our approach computationally more lightweight than related works, which use fully convolutional networks [13,15,20] or geometry-based segmentation algorithms [3,15,17] to optimize the network results. Furthermore, our filter pipeline makes our approach independent of complex trained networks as it applies temporal and spatial filters to the standard detection output of a state-of-the-art network.…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…x = bias 1 z = (z 1 + z 2 )/2 light 1 = bias 2 /cosα light 2 = light 1 − bias 1 y = depth depth = light 2 /tanα (12) In the above formula, z is the height position of the heat source, and z 1 and z 2 are the deviations from the origin of the space coordinates at the heights taken at the two positions. In order to reduce the operation error, the average of the two positions is taken as the height deviation.…”
Section: Coordinate Transformation Mapping In 3d Spacementioning
confidence: 99%
“…Traditionally, using a visible-light binocular camera to reconstruct the target is not possible, because it cannot accurately operate on the abnormal temperature point area [ 7 , 8 , 9 , 10 ]. At present, the most commonly used temperature detection methods use sensor contact measurements [ 11 , 12 , 13 ]. However, there are installation and use problems in engineering applications, so non-contact space measurements can be used to solve the installation problem.…”
Section: Introductionmentioning
confidence: 99%