2014
DOI: 10.1002/cpe.3413
|View full text |Cite
|
Sign up to set email alerts
|

Evaluating automatically parallelized versions of the support vector machine

Abstract: SUMMARYThe Support Vector Machine (SVM) is a supervised learning algorithm used for recognizing patterns in data. It is a very popular technique in Machine Learning and has been successfully used in applications such as image classification, protein classification, and handwriting recognition. However, the computational complexity of the kernelized version of the algorithm grows quadratically with the number of training examples. To tackle this high computational complexity we have developed a directive-based … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
4
0

Year Published

2016
2016
2022
2022

Publication Types

Select...
2
2
2

Relationship

0
6

Authors

Journals

citations
Cited by 9 publications
(4 citation statements)
references
References 46 publications
0
4
0
Order By: Relevance
“…Their algorithm can be improved in the following two aspects: (i) the level of parallelism is low, since only the kernel value computation is parallelized; (ii) it is not scalable to large datasets due to the assumption that the whole kernel matrix can be stored in the CPU memory. Codreanu et al [7] developed a GPU-based SVM training algorithm that requires holding the training dataset in the GPU memory and uses approximation. In this article, we are interested in improving the efficiency and scalability of the SVM training for regression without approximation.…”
Section: Svm Training Using Gpusmentioning
confidence: 99%
See 1 more Smart Citation
“…Their algorithm can be improved in the following two aspects: (i) the level of parallelism is low, since only the kernel value computation is parallelized; (ii) it is not scalable to large datasets due to the assumption that the whole kernel matrix can be stored in the CPU memory. Codreanu et al [7] developed a GPU-based SVM training algorithm that requires holding the training dataset in the GPU memory and uses approximation. In this article, we are interested in improving the efficiency and scalability of the SVM training for regression without approximation.…”
Section: Svm Training Using Gpusmentioning
confidence: 99%
“…We use four real world datasets from the UCI Machine Learning Repository 6 and the LibSVM site 7 . Specifically, the datasets are: (i) Amazon, which contains an anony-mized sample of access provisioned within the Amazon company and person business title serves as the target value of the regression; (ii) CT-slices, which contains features extracted from CT images, and the relative location of the image on the axial axis serves as the target value; (iii) E2006-tfidf, which contains features of corpus and tf-idf of unigrams, and (iv) KDD-CUP'98, which is the dataset used for the KDD Cup 1998 Data Mining Tools Competition, and their first features serve as the target values.…”
Section: Experimental Studymentioning
confidence: 99%
“…Codreanu et al [23], in their paper Evaluating Automatically Parallelized Versions of the Support Vector Machine, deal with the parallelization on multi-core computers of SVM supervised learning algorithm. They propose a new gradient-ascent-based SVM algorithm combined with a particle swarm optimization algorithm for automatic parameter tuning.…”
Section: Pattern Recognitionsmentioning
confidence: 99%
“…ThunderSVM [14] is a rather new SVM implementation supporting CPUs and NVIDIA Graphics Processing Units (GPUs) using the Compute Unified Device Architecture (CUDA). Additionally, there have been various other attempts to port SVMs efficiently to GPUs, e.g., with CUDA [15,16,17], but also using other languages and frameworks such as the Open Computing Language (OpenCL) [18], Python [19,20], OpenACC [21], or SYCL [22]. However, all approaches mentioned so far are based on SMO, an inherently serial algorithm, and, therefore, not very well suited for high performance GPU implementations and vast data sets.…”
Section: Introductionmentioning
confidence: 99%