2019
DOI: 10.1007/978-3-030-29891-3_20
|View full text |Cite
|
Sign up to set email alerts
|

Faster Visual-Based Localization with Mobile-PoseNet

Abstract: Precise and robust localization is of fundamental importance for robots required to carry out autonomous tasks. Above all, in the case of Unmanned Aerial Vehicles (UAVs), efficiency and reliability are critical aspects in developing solutions for localization due to the limited computational capabilities, payload and power constraints. In this work, we leverage novel research in efficient deep neural architectures for the problem of 6 Degrees of Freedom (6-DoF) pose estimation from single RGB camera images. In… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
4
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
4
2

Relationship

0
6

Authors

Journals

citations
Cited by 6 publications
(4 citation statements)
references
References 33 publications
(53 reference statements)
0
4
0
Order By: Relevance
“…Visual-based localization methods can generally be categorized into regression-based, structure-based, and image retrieval-based methods. Regression-based methods include end-to-end visual localization models trained by deep learning that are able to directly obtain the regressed 6DoF camera pose [ 17 , 18 , 19 , 20 ]. However, such methods are not applicable for the visual localization of large-scale scenes and are associated with low accuracies [ 21 ].…”
Section: Related Workmentioning
confidence: 99%
“…Visual-based localization methods can generally be categorized into regression-based, structure-based, and image retrieval-based methods. Regression-based methods include end-to-end visual localization models trained by deep learning that are able to directly obtain the regressed 6DoF camera pose [ 17 , 18 , 19 , 20 ]. However, such methods are not applicable for the visual localization of large-scale scenes and are associated with low accuracies [ 21 ].…”
Section: Related Workmentioning
confidence: 99%
“…The median position/orientation error for the U-SURF descriptors of features sizes 300, 100, 50, 10, 1 along with the state of the art previous work including PoseNet [19], G-Posenet [41], Posenet-U [40], Pose-L [43], Branch-Net [42] Pose-H [52], VidLoc [54], RelocNet [55] and Mobile-PoseNet [56] for the 7 scenes dataset are shown in Table 3. As shown, with 300 U-SURF features, our system displays 0.28 m position error and 9.17 • orientation error.…”
Section: Performance Analysismentioning
confidence: 99%
“…We also extend this work on our locally generated data using a robot mounted with multiple sensors including TABLE 3. The median error in position (m)/ orientation (degrees) for the 7 scenes, with 5 input sizes 300 × 64, 100 × 64, 50 × 64, 10 × 64 and 1 × 64, compared with PoseNet [19], G-Posenet [41], Posenet-U [40], Pose-L [43], BranchNet [42], Pose-H [52], VidLoc [54], RelocNet [55] and Mobile-PoseNet [56]. camera, LIDAR, odometer and SONAR to collect RGB images with ground truth poses.…”
Section: Performance Analysismentioning
confidence: 99%
“…Visual-based localization (VBL) [1][2][3] is one of the promising self-localization technologies that have received growing interest. VBL identifies a device's location in a target space by using cameras to see the device's surroundings, without the dependency on GPS which is designed for outdoor usage or active indoor radio anchor devices which are subject to signal bouncing and interference.…”
Section: Introductionmentioning
confidence: 99%