2010
DOI: 10.1007/s10462-010-9165-y
|View full text |Cite
|
Sign up to set email alerts
|

A review of instance selection methods

Abstract: In supervised learning, a training set providing previously known information is used to classify new instances. Commonly, several instances are stored in the training set but some of them are not useful for classifying therefore it is possible to get acceptable classification rates ignoring non useful cases; this process is known as instance selection. Through instance selection the training set is reduced which allows reducing runtimes in the classification and/or training stages of classifiers. This work is… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
173
0
5

Year Published

2011
2011
2024
2024

Publication Types

Select...
5
3

Relationship

0
8

Authors

Journals

citations
Cited by 336 publications
(179 citation statements)
references
References 44 publications
(63 reference statements)
1
173
0
5
Order By: Relevance
“…There exist generic training set selection techniques [also referred to as instance selection algorithms in the literature (Olvera-López et al 2010)], and those designed for other classifiers [k-nearest neighbors (Angiulli 2005), neural networks (Reeves and Taylor 1998), and many other (Hernandez-Leal et al 2013;Wenyuan et al 2013)], but-due to the specific characteristics of the SVM training process and operation-the majority of SVM training set selection algorithms are crafted for this classifier. In this review, we summarize the state-of-the-art algorithms for selecting SVM training data from large datasets.…”
Section: Motivation and Goalsmentioning
confidence: 99%
“…There exist generic training set selection techniques [also referred to as instance selection algorithms in the literature (Olvera-López et al 2010)], and those designed for other classifiers [k-nearest neighbors (Angiulli 2005), neural networks (Reeves and Taylor 1998), and many other (Hernandez-Leal et al 2013;Wenyuan et al 2013)], but-due to the specific characteristics of the SVM training process and operation-the majority of SVM training set selection algorithms are crafted for this classifier. In this review, we summarize the state-of-the-art algorithms for selecting SVM training data from large datasets.…”
Section: Motivation and Goalsmentioning
confidence: 99%
“… Redundancy: as the name implies is the redundant information such as duplicate instances and derived attributes of others that contain the same information [26], [27]. In Table 3 are shown the approaches to solve the problems related to amount of data and redundancy.…”
Section: Journal Of Computersmentioning
confidence: 99%
“…[26], [27] It is worth mentioning that construction of reports (step of SEMMA methodology) is taking into account in all phases of FDQ-KDT.…”
Section: Journal Of Computersmentioning
confidence: 99%
See 1 more Smart Citation
“…DRTs [18,5,20,9,6,2,13] reduce the computational cost of classification by building a small representative set of the initial training data, called the Condensing Set (CS). The idea behind DRTs is to apply the k-NN classifier over S. Ougiaroglou is supported by a scholarship from the Greek Scholarships Foundations (I.K.Y.…”
Section: Introductionmentioning
confidence: 99%