2013
DOI: 10.1117/12.2008425
|View full text |Cite
|
Sign up to set email alerts
|

Multimedia event detection using visual concept signatures

Abstract: Multimedia Event Detection (MED) is a multimedia retrieval task with the goal of finding videos of a particular event in a large-scale Internet video archive, given example videos and text descriptions. In this paper, we mainly focus on an 'ad-hoc' scenario in MED where we do not use any example video. We aim to retrieve test videos based on their visual semantics using a Visual Concept Signature (VCS) generated for each event only derived from the event description provided as the query. Visual semantics are … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

0
4
0

Year Published

2013
2013
2014
2014

Publication Types

Select...
3
1

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(4 citation statements)
references
References 4 publications
0
4
0
Order By: Relevance
“…In addition, it is also important to legitimately demonstrate the performance in multimedia retrieval by comparing with other related works, which also explore multi-modalities. For example, many researchers explored the cross-modal relationship by applying canonical correlation analysis to better perform in multimedia retrieval (Rasiwasia et al, 2010;Zhang & Liu, 2012), face recognition (Guam, Zhang, Luo, & Lan, 2012), event detection (Younessian, Quinn, Mitamura, & Hauptmann, 2013), etc. In addition, besides positive correlation, Zhai et al also pointed out the importance of capturing negative cross-modality correlation since it can provide exclusive information.…”
Section: Discussionmentioning
confidence: 99%
“…In addition, it is also important to legitimately demonstrate the performance in multimedia retrieval by comparing with other related works, which also explore multi-modalities. For example, many researchers explored the cross-modal relationship by applying canonical correlation analysis to better perform in multimedia retrieval (Rasiwasia et al, 2010;Zhang & Liu, 2012), face recognition (Guam, Zhang, Luo, & Lan, 2012), event detection (Younessian, Quinn, Mitamura, & Hauptmann, 2013), etc. In addition, besides positive correlation, Zhai et al also pointed out the importance of capturing negative cross-modality correlation since it can provide exclusive information.…”
Section: Discussionmentioning
confidence: 99%
“…Particularly, in video retrieval tasks using only text-based query, we can retrieve visual concepts in a video using visual concept signature generated based on a given text query. For instance in [123], visual concept signature idea, explained in Section 6.2.1, was used to tackle the ad-hoc Multimedia Event Detection task. Multimedia Event Detection (MED) is a multimedia retrieval task with the goal of finding videos of a particular event (e.g.…”
Section: Early Fusion Evaluationmentioning
confidence: 99%
“…"getting a vehicle unstuck", "wedding", "making a sandwich", etc) in a large-scale internet video archive, given text descriptions of events. In [123], the test videos were retrieved based on their visual semantics using a Visual Concept Signature (VCS) generated for each event only derived from the event description provided as the query. Visual semantics are described using the Semantic Indexing (SIN) feature which represents the likelihood of predefined visual concepts in a video, similar…”
Section: Early Fusion Evaluationmentioning
confidence: 99%
See 1 more Smart Citation