Proceedings of the 28th ACM International Conference on Multimedia 2020
DOI: 10.1145/3394171.3413896
|View full text |Cite
|
Sign up to set email alerts
|

University-1652: A Multi-view Multi-source Benchmark for Drone-based Geo-localization

Abstract: Figure 1: It is challenging, even for a human, to associate (a) ground-view images with (b) satellite-view images. In this paper, we introduce a new dataset based on the third platform, i.e., drone, to provide real-life viewpoints and intend to bridge the visual gap against views. (c) Here we show two real drone-view images collected from public drone flights on Youtube [1, 8]. (d) In practice, we use the synthetic drone-view camera to simulate the real drone flight. It is based on two concerns. First, the col… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
172
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
2

Relationship

0
6

Authors

Journals

citations
Cited by 181 publications
(172 citation statements)
references
References 40 publications
(119 reference statements)
0
172
0
Order By: Relevance
“…The LCM is compared with the model of Zheng et al (hereinafter referred to as the Zheng model), which is the only study using the University-1652 dataset for cross-view image matching to date [46]. For a fair and reliable comparison, this data division in this study is consistent with the Zheng model.…”
Section: Test Dataset Splitmentioning
confidence: 55%
See 4 more Smart Citations
“…The LCM is compared with the model of Zheng et al (hereinafter referred to as the Zheng model), which is the only study using the University-1652 dataset for cross-view image matching to date [46]. For a fair and reliable comparison, this data division in this study is consistent with the Zheng model.…”
Section: Test Dataset Splitmentioning
confidence: 55%
“…After passing through the convolutional layers of ResNet50, the feature dimension of a 384 × 384 × 3 image is 12 × 12 × 2048. Next, we use the global pooling method to change the image feature into a feature vector with a dimension of 2048 [46]. To change the size of the feature vector, a fully connected layer is added after the pooling layer.…”
Section: Methods Frameworkmentioning
confidence: 99%
See 3 more Smart Citations