2010
DOI: 10.1007/978-3-642-15883-4_14
|View full text |Cite
|
Sign up to set email alerts
|

Learning to Tag from Open Vocabulary Labels

Abstract: Abstract. Most approaches to classifying media content assume a fixed, closed vocabulary of labels. In contrast, we advocate machine learning approaches which take advantage of the millions of free-form tags obtainable via online crowd-sourcing platforms and social tagging websites. The use of such open vocabularies presents learning challenges due to typographical errors, synonymy, and a potentially unbounded set of tag labels. In this work, we present a new approach that organizes these noisy tags into well-… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
24
0

Year Published

2010
2010
2019
2019

Publication Types

Select...
3
2
1

Relationship

0
6

Authors

Journals

citations
Cited by 23 publications
(24 citation statements)
references
References 17 publications
(27 reference statements)
0
24
0
Order By: Relevance
“…Obviously, for a document, its one-hot representation out of the document list is unable to measure similarity between documents and is also subject to the generalization limitation. In recent work of Law, Settles & Mitchell (2010), Latent Direchlet Allocation (LDA) was used to represent terms in form of topics collectively and then a model was trained to map the acoustic content onto the topical representation to facilitate MMIR. Motivated by their work, we employ LDA to represent the local context in our work due to the generality of LDA in representing patterns of collective use.…”
Section: Local Context Representationmentioning
confidence: 99%
See 1 more Smart Citation
“…Obviously, for a document, its one-hot representation out of the document list is unable to measure similarity between documents and is also subject to the generalization limitation. In recent work of Law, Settles & Mitchell (2010), Latent Direchlet Allocation (LDA) was used to represent terms in form of topics collectively and then a model was trained to map the acoustic content onto the topical representation to facilitate MMIR. Motivated by their work, we employ LDA to represent the local context in our work due to the generality of LDA in representing patterns of collective use.…”
Section: Local Context Representationmentioning
confidence: 99%
“…Such semantics provides direct conceptlevel knowledge regarding the concerned multimedia objects. Typical applications include music crowd tagging services (Law, Settles & Mitchell, 2010) and multi-object image data set analysis (Rabinovich et al, 2007). Thanks to crowd-sourced annotation (Turnbull, Barrington & Lanckriet, 2008) and game-based tags collection (Law et al, 2007), large collections of descriptive terms are now accessible.…”
Section: Introductionmentioning
confidence: 99%
“…Finally, some machine learning algorithm is trained over these features in order to obtain a classifier for each tag. Often, the machine learning algorithm attempts to model the semantic relations between the tags [11]. A few state-of-the-art automatic annotation systems are briefly described in section 6.3.…”
Section: Related Approachesmentioning
confidence: 99%
“…A problem with such an approach is the large number of tags which leads to scalability issues. Furthermore, tags can potentially come from an open-vocabulary and be sparse [28]. Another issue could be synonymy where two different tags may have the same meaning.…”
Section: Predicting Tags For Non-tagged Webpagesmentioning
confidence: 99%
“…Another issue could be synonymy where two different tags may have the same meaning. To address these issues in the context of music clip tag prediction, [28] proposed a framework that organizes tags into semantically meaningful classes using topic models, and then predicts these classes given a non-tagged piece of music. Such an approach can be useful for webpage tag prediction as well.…”
Section: Predicting Tags For Non-tagged Webpagesmentioning
confidence: 99%