2016
DOI: 10.1145/2775109
|View full text |Cite
|
Sign up to set email alerts
|

Modality-Dependent Cross-Media Retrieval

Abstract: In this article, we investigate the cross-media retrieval between images and text, that is, using image to search text (I2T) and using text to search images (T2I). Existing cross-media retrieval methods usually learn one couple of projections, by which the original features of images and text can be projected into a common latent space to measure the content similarity. However, using the same projections for the two different retrieval tasks (I2T and T2I) may lead to a tradeoff between their respective perfor… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
34
0
1

Year Published

2017
2017
2022
2022

Publication Types

Select...
6
2

Relationship

0
8

Authors

Journals

citations
Cited by 76 publications
(35 citation statements)
references
References 28 publications
0
34
0
1
Order By: Relevance
“…On NUS-WIDE, the MAP value of our method is far higher compared with the previous and well-known methods. It is about 14.8%, 14.6%, 14.8% higher than that of MDCR for text query image task [19], image query text task and average scores. The reason is that the integration of the predictive labels for the testing images and testing texts can promote each other.…”
Section: Results On the Wikipedia Dataset And Nus-wide Datasetmentioning
confidence: 81%
See 1 more Smart Citation
“…On NUS-WIDE, the MAP value of our method is far higher compared with the previous and well-known methods. It is about 14.8%, 14.6%, 14.8% higher than that of MDCR for text query image task [19], image query text task and average scores. The reason is that the integration of the predictive labels for the testing images and testing texts can promote each other.…”
Section: Results On the Wikipedia Dataset And Nus-wide Datasetmentioning
confidence: 81%
“…MDCR (Modality-dependent cross-media retrieval) [19]: It belongs to Task-specific Cross-modal Retrieval (TSCR). In other words, it uses different mapping mechanisms for different cross-media retrieval tasks.…”
Section: Compared Methods and Evaluation Metricmentioning
confidence: 99%
“…The second is using the same projections for all retrieval tasks (such as image→text and text→image retrieval). Instead, Wei et al [87] propose to learn different projection matrices for image→text retrieval and text→image retrieval. The idea of training different models for different tasks is also presented in the works such as [20].…”
Section: H Other Methodsmentioning
confidence: 99%
“…In a common semantic subspace, data with the same semantics are similar to each other through potential relationships. Wei et al proposed a modality-dependent cross-media retrieval method [40]. e method focuses on the retrieval direction and uses the semantic information of the query modality to project the data into the semantic space of the query modality.…”
Section: Related Workmentioning
confidence: 99%
“…Otherwise, rel k � 0; R k is the number of related items in the top k returns. To evaluate the performance of the proposed GRMD retrieval method, we compare GRMD with the canonical correlation analysis (CCA) [22], kernel canonical correlation analysis (KCCA) [19], semantic matching (SM) [22], semantic correlation matching (SCM) [22], three-view canonical correlation analysis (T-V CCA) [42], generalized multiview linear discriminant analysis (GMLDA) [29], generalized multiview canonical correlation analysis (GMMFA) [29], modalitydependent cross-media retrieval (MDCR) [40], joint feature selection and subspace learning (JFSSL) [43], joint latent subspace learning and regression (JLSLR) [44], generalized semisupervised structured subspace learning (GSSSL) [45], a cross-media retrieval algorithm based on the consistency of collaborative representation (CRCMR) [46], cross-media retrieval based on linear discriminant analysis (CRLDA) [47], and cross-modal online low-rank similarity (CMOLRS) function learning method [48]. e descriptions and characteristics of the above comparison methods used in the whole experiment are summarized in Table 2.…”
Section: Experimental Settingsmentioning
confidence: 99%