2022
DOI: 10.1007/978-3-031-20056-4_8
|View full text |Cite
|
Sign up to set email alerts
|

Demystifying Unsupervised Semantic Correspondence Estimation

Abstract: We explore semantic correspondence estimation through the lens of unsupervised learning. We thoroughly evaluate several recently proposed unsupervised methods across multiple challenging datasets using a standardized evaluation protocol where we vary factors such as the backbone architecture, the pre-training strategy, and the pre-training and finetuning datasets. To better understand the failure modes of these methods, and in order to provide a clearer path for improvement, we provide a new diagnostic framewo… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
7
1

Year Published

2023
2023
2024
2024

Publication Types

Select...
5
1

Relationship

0
6

Authors

Journals

citations
Cited by 7 publications
(8 citation statements)
references
References 65 publications
0
7
1
Order By: Relevance
“…While its unsupervised objective clusters semantically similar features, here we do the opposite and learn features that lead to unique corre-spondences. Aygun et al [2] use an equivariant representation learning approach, inspired by [24,53,54]. Similar to our work, the method learns to project self-supervised ViT features to a semantic correspondence space.…”
Section: Related Workmentioning
confidence: 96%
See 3 more Smart Citations
“…While its unsupervised objective clusters semantically similar features, here we do the opposite and learn features that lead to unique corre-spondences. Aygun et al [2] use an equivariant representation learning approach, inspired by [24,53,54]. Similar to our work, the method learns to project self-supervised ViT features to a semantic correspondence space.…”
Section: Related Workmentioning
confidence: 96%
“…3) to perform best. The La architecture, similar to [2], cannot account for the appearance of the target features by design. The LLA architecture, can do so but learns a degenerate solution: the model can satisfy Eq.…”
Section: Learning a Matcher By Adapting Featuresmentioning
confidence: 99%
See 2 more Smart Citations
“…To overcome the lack of large datasets with ground-truth correspondence, recent work seeks to combine the idea of distilling deep features from a network trained with self-supervision on large-scale image datasets. Some of these works optimize for proxy losses computed with known transformations [4,33,40,58,63,70,73,74,77]. Like these methods, we also train our network to be equivariant to synthetic geometric transformations.…”
Section: Related Workmentioning
confidence: 99%