2019
DOI: 10.3390/rs11192243
|View full text |Cite
|
Sign up to set email alerts
|

AE-GAN-Net: Learning Invariant Feature Descriptor to Match Ground Camera Images and a Large-Scale 3D Image-Based Point Cloud for Outdoor Augmented Reality

Abstract: Establishing the spatial relationship between 2D images captured by real cameras and 3D models of the environment (2D and 3D space) is one way to achieve the virtual–real registration for Augmented Reality (AR) in outdoor environments. In this paper, we propose to match the 2D images captured by real cameras and the rendered images from the 3D image-based point cloud to indirectly establish the spatial relationship between 2D and 3D space. We call these two kinds of images as cross-domain images, because their… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
7
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
1

Relationship

1
5

Authors

Journals

citations
Cited by 8 publications
(7 citation statements)
references
References 50 publications
0
7
0
Order By: Relevance
“…To improve the shortcomings of PPFNet, Deng et al ( 2018a ) further proposed PPF-FoldNet for unsupervised learning of 3D local descriptors in point clouds. In this network, the source point cloud and normal vector are not included in the coding, but the point cloud features are sent to the automatic encoder (AE) like FoldingNet (Hinton and Zemel, 1993 ; Yang et al, 2018 ; Liu et al, 2019a ). After training, the set distance can be used to reconstruct the point pair features.…”
Section: Complete Overlap Point Cloud Registrationmentioning
confidence: 99%
“…To improve the shortcomings of PPFNet, Deng et al ( 2018a ) further proposed PPF-FoldNet for unsupervised learning of 3D local descriptors in point clouds. In this network, the source point cloud and normal vector are not included in the coding, but the point cloud features are sent to the automatic encoder (AE) like FoldingNet (Hinton and Zemel, 1993 ; Yang et al, 2018 ; Liu et al, 2019a ). After training, the set distance can be used to reconstruct the point pair features.…”
Section: Complete Overlap Point Cloud Registrationmentioning
confidence: 99%
“…The most relevant works with cross-domain image (ground camera image and rendered image) patch matching is H-Net [26], H-Net++ [26], SiamAM-Net [1] and AE-GAN-Net [27]. H-Net only performs the binary matching judgments Final triplet of cross-domain image patches.…”
Section: Related Workmentioning
confidence: 99%
“…When capturing camera images, we set up the mobile phone on a handheld gimbal to acquire a more accurate camera pose, which reduces the camera jitter, to capture images. In addition, the same as described in the literature [1] and [27], to better obtain the cross-domain image patch dataset, we have carried out manual supervision, which the ground camera image and rendered image pairs with obvious deviations are discarded.…”
Section: A Datasetmentioning
confidence: 99%
See 1 more Smart Citation
“…AR can enable users to experience the real world in which virtual objects and real objects coexist, and interact with them in the real time. In the past two decades, AR application has been a trending research topic in many areas, such as education, entertainment, medicine, and industry [2,3]. Volkswagen intended to use AR to compare the calculated crash test imagery with the actual case [4].…”
Section: Introductionmentioning
confidence: 99%