Proceedings of the 2022 International Conference on Multimedia Retrieval 2022
DOI: 10.1145/3512527.3531369
|View full text |Cite
|
Sign up to set email alerts
|

Efficient Linear Attention for Fast and Accurate Keypoint Matching

Abstract: Figure 1: Our method versus the bigger SOTAs-SuperGlue and SGMNet-on speed (left), image matching * (top), and 3D reconstruction (bottom).

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
4
1

Relationship

0
5

Authors

Journals

citations
Cited by 7 publications
(2 citation statements)
references
References 32 publications
0
2
0
Order By: Relevance
“…With the advent of the deep learning era, the robustness of local features under illumination variation and viewpoint change can be significantly improved by learning-based methods, such as SuperPoint [27], D2-Net [28], R2D2 [29], and AWDesc [30]. In addition to learning-based detector-descriptors, some works recently focus on learning local feature matching [6], [20], [21], [23], [24], [31]. As a pioneering work, SuperGlue [6] requires two sets of keypoints and their visual descriptors as input and learns their correspondences using an attentional GNN with complete graphs over keypoints within and across images.…”
Section: Previous Arts a Local Feature Matchingmentioning
confidence: 99%
See 1 more Smart Citation
“…With the advent of the deep learning era, the robustness of local features under illumination variation and viewpoint change can be significantly improved by learning-based methods, such as SuperPoint [27], D2-Net [28], R2D2 [29], and AWDesc [30]. In addition to learning-based detector-descriptors, some works recently focus on learning local feature matching [6], [20], [21], [23], [24], [31]. As a pioneering work, SuperGlue [6] requires two sets of keypoints and their visual descriptors as input and learns their correspondences using an attentional GNN with complete graphs over keypoints within and across images.…”
Section: Previous Arts a Local Feature Matchingmentioning
confidence: 99%
“…Despite the inspiring progress, these methods are inherently limited by the following two aspects: 1) Their inputs are putative correspondences provided by an NN search, so it is impossible to retrieve correct matches beyond these ones, which means that the upper bound on the performance is limited by the quality of initial matches; 2) Local visual descriptors and correspondence coordinates are separately used for match generation and mismatch rejection, respectively, neglecting the interaction between visual information from high-level image representation and geometric information from 2D keypoint distribution, calling for the research towards the design of effective feature matching methods to break the limitation of vanilla nearest neighbor correspondences. To this end, a line of research directly learns to find the partial assignment between two sets of local features with an attention-based Graph Neural Network (GNN) [20], [21]. Representatively, SuperGlue [6] constructs densely-connected graphs over intra-and inter-image keypoints and performs message passing by means of self-and cross-attention.…”
Section: Introductionmentioning
confidence: 99%