Proceedings of the 2016 ACM on International Conference on Multimedia Retrieval 2016
DOI: 10.1145/2911996.2912000
|View full text |Cite
|
Sign up to set email alerts
|

Correlation Autoencoder Hashing for Supervised Cross-Modal Search

Abstract: Due to its storage and query efficiency, hashing has been widely applied to approximate nearest neighbor search from large-scale datasets. While there is increasing interest in cross-modal hashing which facilitates cross-media retrieval by embedding data from different modalities into a common Hamming space, how to distill the cross-modal correlation structure effectively remains a challenging problem. In this paper, we propose a novel supervised cross-modal hashing method, Correlation Autoencoder Hashing (CAH… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
41
0
1

Year Published

2018
2018
2023
2023

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 90 publications
(42 citation statements)
references
References 28 publications
0
41
0
1
Order By: Relevance
“…This may be due to several reasons. First, the method of CAH [17] still uses handcrafted image features as the input to their deep neural networks whereas our model starts learning from raw images. In the methods of DVSH [13] and CMDVH [14], the modality-specific hash functions are learned such that the non-linear relationship of samples from different modalities are exploited.…”
Section: ) Comparison To Deep Cross-modal Hashing Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…This may be due to several reasons. First, the method of CAH [17] still uses handcrafted image features as the input to their deep neural networks whereas our model starts learning from raw images. In the methods of DVSH [13] and CMDVH [14], the modality-specific hash functions are learned such that the non-linear relationship of samples from different modalities are exploited.…”
Section: ) Comparison To Deep Cross-modal Hashing Methodsmentioning
confidence: 99%
“…• CMNN [15]: The method is to learn a similarity preserving network for cross-modalities through a coupled Siamese network with hinge loss. • CAH [17]: A model that is designed with a stacked autoencoder architecture to jointly maximize the feature and semantic correlation across modalities. • DCMH [18]: It is an end-to-end deep learning framework with a negative log likelihood criterion to preserve the similarity between real-value representations having the same class.…”
Section: B Competitors and Evaluation Setupmentioning
confidence: 99%
“…Supervised hashing methods [2,4,14,16,39,40,44] aim to exploit available supervised information (such as labels or the semantic affinities of training data) to improve performance. Brostein et al [2] present a cross-modal hashing approach by preserving the intra-class similarity via eigendecomposition and boosting.…”
Section: Related Workmentioning
confidence: 99%
“…In fact, in standard cross-modal benchmark datasets such as NUS-WIDE [6] and Microsoft COCO [15], an image instance can be assigned to multiple category labels [27], which is beneficial as it permits semantic relevance to be described more accurately across different modalities. Second, these methods enforce a narrowing of the modality gap by constraining the corresponding hash codes with certain pre-defined loss functions [4]. The code length is usually text A collection of baseball players who are at home plate.…”
Section: Introductionmentioning
confidence: 99%
“…Deep cross-modal hashing (DCMH) [24] combines hashing learning and deep feature learning by preserving the semantic similarity between modalities. Correlation autoencoder hashing (CAH) [28] embeds the maximum cross-modal similarity into hash codes using nonlinear deep autoencoders. Correlation hashing network (CHN) [29] jointly learns image and text representations tailored to hash coding and formally controls the quantization error.…”
Section: Related Workmentioning
confidence: 99%