2007
DOI: 10.1109/tpami.2007.1097
|View full text |Cite
|
Sign up to set email alerts
|

Modeling Semantic Aspects for Cross-Media Image Indexing

Abstract: To go beyond the query-by-example paradigm in image retrieval, there is a need for semantic indexing of large image collections for intuitive text-based image search. Different models have been proposed to learn the dependencies between the visual content of an image set and the associated text captions, then allowing for the automatic creation of semantic indices for unannotated images. The task, however, remains unsolved. In this paper, we present three alternatives to learn a Probabilistic Latent Semantic A… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

1
164
1

Year Published

2009
2009
2021
2021

Publication Types

Select...
6
2
1

Relationship

0
9

Authors

Journals

citations
Cited by 161 publications
(166 citation statements)
references
References 34 publications
1
164
1
Order By: Relevance
“…Wu et al [33] proposed more accurate annotations by calculating the contribution of each word to its visibility model. Monay and Gatica-Perez [21] introduced a latent topic model, probabilistic Latent Semantic Analysis (pLSA), into an image annotation algorithm and trained two pLSA models based on the visual features and associated text, respectively. This method then merged the topics in both models by assigning weights according to the entropy of the visual word distribution.…”
Section: Related Workmentioning
confidence: 99%
“…Wu et al [33] proposed more accurate annotations by calculating the contribution of each word to its visibility model. Monay and Gatica-Perez [21] introduced a latent topic model, probabilistic Latent Semantic Analysis (pLSA), into an image annotation algorithm and trained two pLSA models based on the visual features and associated text, respectively. This method then merged the topics in both models by assigning weights according to the entropy of the visual word distribution.…”
Section: Related Workmentioning
confidence: 99%
“…A number of researchers have studied the problem of automatic image annotation in recent years [2][3][4][5][6]1]. Given cluttered images of multiple objects paired with noisy captions, these systems can learn meaningful correspondences between caption words and appearance models.…”
Section: Related Workmentioning
confidence: 99%
“…Most of these automatic annotation systems do not focus on learning such feature configurations. Often, appearance is modeled as a mixture of features (e.g., [5,3,6]) in which common part configurations are reflected in co-occurrence statistics but without spatial information. Similarly, the Markov random field model proposed by Carbonetto et al [4] can represent adjacency relationships but not spatial configurations.…”
Section: Related Workmentioning
confidence: 99%
“…As the discriminative method, we use Support Vector Machine (SVM). As the generative method, we use probabilistic latent topic mixture models [9].…”
Section: Image Classificationmentioning
confidence: 99%