2014 IEEE Congress on Evolutionary Computation (CEC) 2014
DOI: 10.1109/cec.2014.6900490
|View full text |Cite
|
Sign up to set email alerts
|

A combined MapReduce-windowing two-level parallel scheme for evolutionary prototype generation

Abstract: Evolutionary prototype generation techniques have demonstrated their usefulness to improve the capabilities of the nearest neighbor classifier. They act as data reduction algorithms by generating representative points of a given problem. Their main purposes are to speed up the classification process and to reduce the storage requirements and sensitivity to noise of the nearest neighbor rule. Nowadays, with the increment of available data, the use of this kind of reduction techniques becomes more important. How… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
5
0

Year Published

2015
2015
2022
2022

Publication Types

Select...
3
2
1

Relationship

2
4

Authors

Journals

citations
Cited by 6 publications
(5 citation statements)
references
References 20 publications
0
5
0
Order By: Relevance
“…The reduce phase offers several ways to aggregate the partial instance sets, either by: concatenating all partial results (baseline), filtering noisy prototypes, or merging redundant samples. An extension to this method was proposed in [98]. A second phase of parallelization based on windowing is included in this extension on the mappers side.…”
Section: Instance Reductionmentioning
confidence: 99%
“…The reduce phase offers several ways to aggregate the partial instance sets, either by: concatenating all partial results (baseline), filtering noisy prototypes, or merging redundant samples. An extension to this method was proposed in [98]. A second phase of parallelization based on windowing is included in this extension on the mappers side.…”
Section: Instance Reductionmentioning
confidence: 99%
“…Current research is typically focused on the use of divide-and-conquer approaches, implemented with big data technologies, to parallelise the execution of instance reduction approaches. We can also find an approximation strategy, called windowing [31], which estimates the fitness value of RS using a random subset of T R at every iteration of the search (this reduces significantly the cost, but could mislead the search). However, the use of more sophisticated surrogate models to reduce the number of evaluations for instance reduction algorithms has been neglected.…”
Section: Computational Cost Of the Objective Functionmentioning
confidence: 99%
“…This approach is different from the partitioning of the data considered in MapReduce because all the information of the training set is available, but only a part of it is used in each iteration. This technique has already been extended to other evolutionary algorithms used in data mining tasks [15].…”
Section: B a Two-level Parallelization Model For Eusmentioning
confidence: 99%
“…To do so, we consider the MapReduce algorithm, which splits the data in different chunks that are processed in different computing nodes (mappers), such that EUS can be applied concurrently. Moreover, the time required by EUS is further reduced by considering a windowing scheme inside the algorithm [14], [15]. This is specifically designed for the class imbalance problem.…”
Section: Introductionmentioning
confidence: 99%