Proceedings of the 15th ACM International Conference on Multimedia 2007
DOI: 10.1145/1291233.1291295
|View full text |Cite
|
Sign up to set email alerts
|

Segregated feedback with performance-based adaptive sampling for interactive news video retrieval

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
14
0

Year Published

2008
2008
2012
2012

Publication Types

Select...
3
2
2

Relationship

2
5

Authors

Journals

citations
Cited by 13 publications
(14 citation statements)
references
References 3 publications
0
14
0
Order By: Relevance
“…It utilizes an extensive annotation strategy for interactive search based on human reaction time. The VisionGo system [15]) supports multi-modal query and combines text derived from ASR, semantic concept classification, and low-level visual features and motion to rank videos. An active learning strategy is also used to maximize the users' interaction efforts.…”
Section: Query-by-concept Video Search Systemsmentioning
confidence: 99%
“…It utilizes an extensive annotation strategy for interactive search based on human reaction time. The VisionGo system [15]) supports multi-modal query and combines text derived from ASR, semantic concept classification, and low-level visual features and motion to rank videos. An active learning strategy is also used to maximize the users' interaction efforts.…”
Section: Query-by-concept Video Search Systemsmentioning
confidence: 99%
“…For example, Luan et al [269] iteratively select videos that are the most relevant to the query until the number of videos labeled as relevant by users in an iteration step becomes very small. Then, the videos closest to the classifier boundary are returned to users for identification and the system is updated using the identified videos.…”
Section: ) Explicit Relevance Feedbackmentioning
confidence: 99%
“…One problem concerning the usage of too many visual features is the curse of high dimensionality, which may degrade performance and introduce large time cost. In order to cater to real-time retrieval, we restrict the feature size to a 116-feature vector for each key frame, consisting of 27-dimension color moment features (including 1st, 2nd, and 3rd moments) obtained at a 3 Â 3 block, 80-dimension normalized local edge histogram texture feature, eight directional motion features and one global motion feature [25].…”
Section: Low-level Visual Featuresmentioning
confidence: 99%