2015
DOI: 10.1007/s10479-015-1956-8
|View full text |Cite
|
Sign up to set email alerts
|

Nearest neighbors methods for support vector machines

Abstract: A key issue in the practical applicability of the support vector machine methodology is the identification of the support vectors in very large data sets, a problem to which a great deal of attention has been given in the literature. In the present article we propose methods based on sampling and nearest neighbors, that allow for an efficient implementation of an approximate solution to the classification problem and, at least in some problems, will help in identifying a significant fraction of the support vec… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
15
0

Year Published

2019
2019
2025
2025

Publication Types

Select...
6

Relationship

0
6

Authors

Journals

citations
Cited by 6 publications
(15 citation statements)
references
References 25 publications
0
15
0
Order By: Relevance
“…A different approach is presented in [14] where the reduction of the data includes the assignments of large weights to important samples and the reduction of the features, by using graph and self-paced learning. The methods in [6] and [5] use nearest neighbors with sub-sampling in order to select a subset of significant instances. They start by training an SVM formed by a very small sub-samples of the data set.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…A different approach is presented in [14] where the reduction of the data includes the assignments of large weights to important samples and the reduction of the features, by using graph and self-paced learning. The methods in [6] and [5] use nearest neighbors with sub-sampling in order to select a subset of significant instances. They start by training an SVM formed by a very small sub-samples of the data set.…”
Section: Related Workmentioning
confidence: 99%
“…And, if λ is chosen smaller than 1, the contribution of λ n A n X is small for large n. Then, in practice it is enough to sum over a finite number N k of terms. However, even with these considerations, the running time for the computation of the Kernel using (6) is O(k 6 ) with k = |V ||V ′ |, so it can be very high for large graphs. A cheaper way to compute the Random Walk Kernel is to use the following equivalence…”
Section: Basic Facts and Notationmentioning
confidence: 99%
“…Another contribution of the present work is to propose a new subsampling algorithm by improving the results of Camelo et al (2015) [4], at least in a significant number of cases, by enriching the subsample with more candidates to support vectors using bagging and importance sampling. This is achieved by looking simultaneously at different samples and searching for neighbors according to the candidates' intensity.…”
Section: Our Contributionmentioning
confidence: 99%
“…By testing on benchmark examples and comparing with state-of-the-art methodologies (such as the ones proposed in [4], LibSVM [5], SVM light [6], and decision trees [7]), we show that our proposed method achieves a fast solution to the training SVM problem without a significant loss in the performance accuracy. It is important to highlight that one goal of this paper is to compare algorithms using the same working framework in order to conclude about efficiency and effectiveness.…”
Section: Our Contributionmentioning
confidence: 99%
“…For example, specialized algorithms for solving quadratic programming have been suggested, including the sequential minimal optimization (Platt, 1998) and various decomposition methods used in the LibLinear software library (Hsieh et al, 2008). Other fast computation methods based on low-rank approximation (Williams and Seeger, 2000), gradient descent (Bordes et al, 2005;Shalev-Shwartz et al, 2011;Wang et al, 2012), core set (Tsang et al, 2005), and nearest neighbor (Camelo et al, 2015) have also been developed. However, it is worth noting that most of these methods still incur a computational cost of at least O(N 2 ) or lack optimal statistical guarantees.…”
Section: Introductionmentioning
confidence: 99%