2021
DOI: 10.1016/j.knosys.2020.106582
|View full text |Cite
|
Sign up to set email alerts
|

A fast parallel attribute reduction algorithm using Apache Spark

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
1
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
6
1
1

Relationship

1
7

Authors

Journals

citations
Cited by 15 publications
(3 citation statements)
references
References 24 publications
0
1
0
Order By: Relevance
“…Spark is a distributed computing framework based on in-memory computing with faster computation speed and better iterative performance than Hadoop MapReduce [31]. Yin et al [32] designed a new parallel attribute reduction4 based on Spark to resolve the limitations of MapReduce. Luo et al [33], [34] proposed a novel Spark parallel attribute reduction based on a rough hypercuboid model, which employed two parallel strategies, vertical partitioning and horizontal partitioning.…”
Section: A Novel Spark-based Attribute Reduction and Neighborhood Cla...mentioning
confidence: 99%
“…Spark is a distributed computing framework based on in-memory computing with faster computation speed and better iterative performance than Hadoop MapReduce [31]. Yin et al [32] designed a new parallel attribute reduction4 based on Spark to resolve the limitations of MapReduce. Luo et al [33], [34] proposed a novel Spark parallel attribute reduction based on a rough hypercuboid model, which employed two parallel strategies, vertical partitioning and horizontal partitioning.…”
Section: A Novel Spark-based Attribute Reduction and Neighborhood Cla...mentioning
confidence: 99%
“…In this section, we use speedup, scaleup, and sizeup [30] to evaluate the parallel performance of the proposed algorithm and compare it with the standard Spark-MLRF algorithm.…”
Section: Parallel Performance Evaluationmentioning
confidence: 99%
“…Attribute reduction [4,30,34,55] is an essential idea of rough set theory, it can select attributes or features with the highest discriminative or predictive power from the original data to reduce data redundancy or noise [20,36,54]. In the era of big data, attribute reduction can eliminate irrelevant or unimportant attributes from high-dimensional datasets [7,56], improve data quality, discover the potential value and knowledge of data, reduce the storage and computation costs of data, and improve the speed of data processing [8,57].…”
Section: Introductionmentioning
confidence: 99%