2019 International Conference on Robotics and Automation (ICRA) 2019
DOI: 10.1109/icra.2019.8794448
|View full text |Cite
|
Sign up to set email alerts
|

Multi-Task Template Matching for Object Detection, Segmentation and Pose Estimation Using Depth Images

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
24
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
6
2

Relationship

1
7

Authors

Journals

citations
Cited by 44 publications
(24 citation statements)
references
References 19 publications
0
24
0
Order By: Relevance
“…Park et al proposed an MTTM method for object detection, segmentation, and pose estimation. In contrast to other methods that rely on the color and texture information, the MTTM method uses depth images as input and estimates the object poses using the nearest neighborhood matching [8].…”
Section: B Geometry-based Methodsmentioning
confidence: 99%
See 2 more Smart Citations
“…Park et al proposed an MTTM method for object detection, segmentation, and pose estimation. In contrast to other methods that rely on the color and texture information, the MTTM method uses depth images as input and estimates the object poses using the nearest neighborhood matching [8].…”
Section: B Geometry-based Methodsmentioning
confidence: 99%
“…Thanks to the well-developed deep learning-based object detection methods, the view-based 6D target object detection methods such as [5], [6] and the view-geometry fusion method such as [7] yielded promising results on a public dataset. Similar to the RGB images, researchers represented the object geometry with the depth image and achieved satisfactory results [8].…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…Our approach for grasping relies on first finding the nearest observation in a database and second predicting dense geometric correspondences to transform a successful grasp pose to a new observation. Retrieving similar samples has been addressed using learned feature descriptors from RGB-D images (Wohlhart and Lepetit, 2015 ; Balntas et al, 2017 ; Park et al, 2019a ). These employ the triplet loss to train a network to produce smaller feature distances for pairs of images with similar viewpoints while producing larger feature distances for pairs of images with different viewpoints or those that contain different object classes.…”
Section: Related Workmentioning
confidence: 99%
“…It features the advantages of improved efficiency and accuracy respectively in terms of learning and prediction due to the fact that commonalities and differences across tasks are exploited [33]. In addition, it is effective in the avoidance of overfitting on a specific task since the network model is regularized [34]. Wang et al managed to improve the 6D object pose estimation performance, especially under the condition of occlusion via a multitask learning network combining object recognition with pose estimation [35].…”
Section: Multitask Learningmentioning
confidence: 99%