2011 Eighth International Conference on Fuzzy Systems and Knowledge Discovery (FSKD) 2011
DOI: 10.1109/fskd.2011.6020072
|View full text |Cite
|
Sign up to set email alerts
|

A distributed SVM for scalable image annotation

Abstract: Machine learning techniques have facilitated image retrieval by automatically classifying and annotating images with keywords. Among them Support Vector Machines (SVMs) are used extensively due to their generalization properties. However, SVM training is notably a computationally intensive process especially when the training dataset is large. This paper presents MRSVM, a distributed SVM algorithm for large scale image annotation which partitions the training data set into smaller subsets and train SVM in para… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
8
0

Year Published

2011
2011
2022
2022

Publication Types

Select...
3
2
1

Relationship

1
5

Authors

Journals

citations
Cited by 15 publications
(8 citation statements)
references
References 14 publications
0
8
0
Order By: Relevance
“…Considering the total processing power of the cluster is P ¼ X n i¼1 p i where n represents the number of the processors employed in the cluster and p i represents the processing speed of the ith processor, for a Hadoop cluster with a total computing capacity P, the level of heterogeneity of the Hadoop cluster can be defined using Eq. (5).…”
Section: Load Balancingmentioning
confidence: 99%
See 1 more Smart Citation
“…Considering the total processing power of the cluster is P ¼ X n i¼1 p i where n represents the number of the processors employed in the cluster and p i represents the processing speed of the ith processor, for a Hadoop cluster with a total computing capacity P, the level of heterogeneity of the Hadoop cluster can be defined using Eq. (5).…”
Section: Load Balancingmentioning
confidence: 99%
“…However, with the rapid development of a variety of computer systems in the electric power grid, it has become a challenging issue to ensure a large amount of CIM data to be correct and consistent all the time.MapReduce has become a major computing model in support of data-intensive applications [4]. MapReduce facilitates a number of important functions such as partitioning the input data, scheduling MapReduce jobs across a cluster of participating nodes, handling node failures, and managing the required network communications [5]. We have implemented a MapReduce-based parallel K-means clustering for scalable information retrieval [6].…”
mentioning
confidence: 99%
“…Catanzaro et al proposed a parallel SMO algorithm based on MapReduce, but this algorithm is somewhat inefficient because of the iterative nature of SMO. In , Alham discussed an efficient and scalable approach based on a single MapReduce phase. They announced that this method has minimal data movement between nodes and that it also minimizes communication overheads.…”
Section: Multimedia Applications In Mapreducementioning
confidence: 99%
“…SMO speeds up the training phase only, with no control over the number of support vectors or testing time. To achieve additional acceleration, many parallel implementations of SMO (Zeng et al 2008;Peng, Ma, and Hong 2009;Catanzaro et al 2008;Alham et al 2010;Cao et al 2006) were developed on various parallel programming platforms, including graphics processing unit (GPU) (Catanzaro et al 2008), Hadoop MapReduce (Alham et al 2010), and message passing interface (MPI) (Cao et al 2006).…”
Section: Improving Svm Computational Requirementsmentioning
confidence: 99%