Proceedings of the Sixth ACM SIGKDD International Conference on Knowledge Discovery and Data Mining 2000
DOI: 10.1145/347090.347155
|View full text |Cite
|
Sign up to set email alerts
|

Towards scalable support vector machines using squashing

Abstract: Support vector machines (SVMs) provide classi cation models with strong theoretical foundations as well as excellent empirical performance on a variety of applications. One of the major drawbacks of SVMs is the necessity to solve a large-scale quadratic programming problem. This paper combines likelihood-based squashing with a probabilistic formulation of SVMs, enabling fast training on squashed data sets. We reduce the problem of training the SVMs on the weighted \squashed" data to a quadratic programming pro… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
35
0

Year Published

2002
2002
2019
2019

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 42 publications
(35 citation statements)
references
References 6 publications
(4 reference statements)
0
35
0
Order By: Relevance
“…Pavlov applied data squashing to support vector machine: a classifier which maximizes margins of training examples under a similar philosophy to boosting [9]. Nakayasu substituted a product-sum matrix for the CF vector, and applied their method to Bayesian classification [8].…”
Section: Application Of Data Squashing To Classification and Regressionmentioning
confidence: 99%
“…Pavlov applied data squashing to support vector machine: a classifier which maximizes margins of training examples under a similar philosophy to boosting [9]. Nakayasu substituted a product-sum matrix for the CF vector, and applied their method to Bayesian classification [8].…”
Section: Application Of Data Squashing To Classification and Regressionmentioning
confidence: 99%
“…For example, it has been proposed to construct an approximate SVM by approximating the Gram matrix with a smaller matrix using either low rank representation [12] or sampling [1,29]. Assuming a linear kernel is used, Pavlov et al [23] proposed to squash the original training dataset into a limited number of representatives and construct an approximate SVM using these representatives. Although these algorithms, together with many other algorithms for approximate SVMs, are well motivated and have been shown to be very effective experimentally, there is no direct theoretical justification on the generalization performance of the resulting approximate SVMs.…”
Section: Introductionmentioning
confidence: 99%
“…Enhancing the SVM training process with clustering or similar techniques has been examined with several variations in [34], [26] and [29]. Based on a hierarchical micro-clustering algorithm, Yu et al [34] proposed a scalable algorithm to train support vector machines with a linear kernel.…”
Section: Introductionmentioning
confidence: 99%
“…In [29], Shih et al proposed a technique called text bundling, where the training data are partitioned into clusters based on a Rocchio score, and data within a cluster are replaced by their mean. Pavlov et al [26] proposed a strategy to speed up the training process by squashing. Both [29] and [26] used only linear kernels in their experiments.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation