2017
DOI: 10.1371/journal.pone.0178161
|View full text |Cite
|
Sign up to set email alerts
|

Distributed optimization of multi-class SVMs

Abstract: Training of one-vs.-rest SVMs can be parallelized over the number of classes in a straight forward way. Given enough computational resources, one-vs.-rest SVMs can thus be trained on data involving a large number of classes. The same cannot be stated, however, for the so-called all-in-one SVMs, which require solving a quadratic program of size quadratically in the number of classes. We develop distributed algorithms for two all-in-one SVM formulations (Lee et al. and Weston and Watkins) that parallelize the co… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
9
0

Year Published

2018
2018
2022
2022

Publication Types

Select...
5
3

Relationship

1
7

Authors

Journals

citations
Cited by 12 publications
(10 citation statements)
references
References 34 publications
1
9
0
Order By: Relevance
“…SVMs not only perform binary classification, they are also able to do multiclass classification. Four such models are: all-vs-all (AVA) SVM, one-vs-all (OVA) SVM, structured SVM [43], and the Weston and Watkins version [44]. Besides linear classification, SVMs can perform non-linear classification.…”
Section: Combining Classification and Regression Tasks Support Vectormentioning
confidence: 99%
“…SVMs not only perform binary classification, they are also able to do multiclass classification. Four such models are: all-vs-all (AVA) SVM, one-vs-all (OVA) SVM, structured SVM [43], and the Weston and Watkins version [44]. Besides linear classification, SVMs can perform non-linear classification.…”
Section: Combining Classification and Regression Tasks Support Vectormentioning
confidence: 99%
“…Existing data-dependent analyses build on either the structural result (1) or (3), which either ignores the correlation among predictors associated to individual class labels or requires f i to be Lipschitz continuous w.r.t. the 2 -norm.…”
Section: Data-dependent Bounds By Gaussian Complexitiesmentioning
confidence: 99%
“…Motivated by the ambiguity in class labels caused by the rapid increase in number of classes in modern computer vision benchmarks, Lapin et al [43,44] introduce the top-k MC-SVM by using the topk hinge loss to allow k predictions for each object x. For any t ∈ R c , let the bracket [·] denote a permutation such that [j] is the index of the j-th largest score, i.e., t [1]…”
Section: Top-k Mc-svmmentioning
confidence: 99%
See 1 more Smart Citation
“…A easy way is to divide a multiclassification problem into many binary classification problems. Various approaches were proposed to achieve efficient multi classificaton [5,6,7]. The one-aganinst-one (OAO) with a voting strategy [8] and one-against-rest (OAR or one-against-all) with a winer-take-all strategy [9] are two popular methods for doing this dividation.…”
Section: Introductionmentioning
confidence: 99%