Proceedings of the Thirtieth International Joint Conference on Artificial Intelligence 2021
DOI: 10.24963/ijcai.2021/156
|View full text |Cite
|
Sign up to set email alerts
|

Deep Unified Cross-Modality Hashing by Pairwise Data Alignment

Abstract: With the increasing amount of multimedia data, cross-modality hashing has made great progress as it achieves sub-linear search time and low memory space. However, due to the huge discrepancy between different modalities, most existing cross-modality hashing methods cannot learn unified hash codes and functions for modalities at the same time. The gap between separated hash codes and functions further leads to bad search performance. In this paper, to address the issues above, we propose a novel end-to-end Deep… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
10
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
3

Relationship

0
7

Authors

Journals

citations
Cited by 11 publications
(10 citation statements)
references
References 1 publication
0
10
0
Order By: Relevance
“…To bridge modality gap, JDSH [8] exploits a joint-modal similarity matrix, while DUCMH [9] relies on data alignment and image-text data pairs. DGCPN [10] explores intrinsic semantic relationships with graph-neighbor coherence to avoid suboptimal retrieval Hamming space. With introducing knowledge distillation scheme, KDCMH [11] trains an unsupervised method as the teacher model used to provide distillation information to guide supervised method.…”
Section: A Non-continuous Cross-modal Hashingmentioning
confidence: 99%
See 1 more Smart Citation
“…To bridge modality gap, JDSH [8] exploits a joint-modal similarity matrix, while DUCMH [9] relies on data alignment and image-text data pairs. DGCPN [10] explores intrinsic semantic relationships with graph-neighbor coherence to avoid suboptimal retrieval Hamming space. With introducing knowledge distillation scheme, KDCMH [11] trains an unsupervised method as the teacher model used to provide distillation information to guide supervised method.…”
Section: A Non-continuous Cross-modal Hashingmentioning
confidence: 99%
“…Hash lookup, a widely used retrieval protocol in hashingbased retrieval [10], considers the retrieved instance whose Hamming distance to the query is less than Hamming radius as a positive sample. When measuring the precision of hash lookup protocol, we hope that a good hashing method can retrieve as many positive samples as possible, i.e., when the query instance o i and the retrieved instance o j are true relevant, the probability of being judged as relevant should be as large as possible, denoted as:…”
Section: Multi-label Semantic Similaritymentioning
confidence: 99%
“…Hashing (Wang et al 2018a;Gao et al 2023;Kou et al 2022;Wang et al 2021) has gained increasing interests in many large-scale applications. Its primary goal is to encode • We propose a novel distributed manifold hashing (DMH) method for compact image set representation.…”
Section: Introductionmentioning
confidence: 99%
“…Hashing Hashing (Wang et al 2018a,a;Gao et al 2023;Kou et al 2022;Wang et al 2021) aims to learn compact hash code in Hamming space while preserving similarity, which can be particularly useful for efficient largescale search. Many existing hashing methods (Gong et al 2013;Shen et al 2015) employ machine learning technique to learn hash function and hash code.…”
Section: Introductionmentioning
confidence: 99%
“…As shown in Figure 1(a), only a short video segment semantically matches the query, while most of the video contents are queryirrelevant. Clearly, TSG tries to break through the barrier between computer vision and natural language processing techniques for more challenging cross-modal grounding (Li et al, ,a, 2022Wang and Shi, 2023;Wang et al, 2021aWang et al, , 2020c.…”
Section: Introductionmentioning
confidence: 99%