2013
DOI: 10.3991/ijoe.v9i3.2765
|View full text |Cite
|
Sign up to set email alerts
|

Big Data Cleaning Algorithms in Cloud Computing

Abstract: Abstract-Big data cleaning is one of the important research issues in cloud computing theory. The existing data cleaning algorithms assume all the data can be loaded into the main memory at one-time, which are infeasible for big data. To this end, based on the knowledge base, a data cleaning algorithm is proposed in cloud computing by Map-Reduce. It extracts atomic knowledge of the selected nodes firstly, then analyzes their relations, deletes the same objects, builds an atomic knowledge sequence based on weig… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2

Citation Types

0
11
0

Year Published

2015
2015
2024
2024

Publication Types

Select...
6
1
1

Relationship

2
6

Authors

Journals

citations
Cited by 16 publications
(11 citation statements)
references
References 9 publications
0
11
0
Order By: Relevance
“…16,17 Sansom, 9 Nguyen K et al, 18 Price D et al, 19 analyzed theoretically how memory residency (the ratio between live memory and the total heap) affects the performance of both collectors, and adopted a dual-mode collector, which switches between copying and compacting according to the memory usage. 16,17 Sansom, 9 Nguyen K et al, 18 Price D et al, 19 analyzed theoretically how memory residency (the ratio between live memory and the total heap) affects the performance of both collectors, and adopted a dual-mode collector, which switches between copying and compacting according to the memory usage.…”
Section: Space Efficient Collectorsmentioning
confidence: 99%
See 1 more Smart Citation
“…16,17 Sansom, 9 Nguyen K et al, 18 Price D et al, 19 analyzed theoretically how memory residency (the ratio between live memory and the total heap) affects the performance of both collectors, and adopted a dual-mode collector, which switches between copying and compacting according to the memory usage. 16,17 Sansom, 9 Nguyen K et al, 18 Price D et al, 19 analyzed theoretically how memory residency (the ratio between live memory and the total heap) affects the performance of both collectors, and adopted a dual-mode collector, which switches between copying and compacting according to the memory usage.…”
Section: Space Efficient Collectorsmentioning
confidence: 99%
“…As copying-based collectors need to reserve half of the available heap, whereas compacting-based collectors do not; for any particular program, there must be a point in the time-space tradeoff where the two kinds of collectors achieve the same performance. 16,17 Sansom, 9 Nguyen K et al, 18 Price D et al, 19 analyzed theoretically how memory residency (the ratio between live memory and the total heap) affects the performance of both collectors, and adopted a dual-mode collector, which switches between copying and compacting according to the memory usage. 20,21 His algorithm works well and its copying space can make use of our algorithm to further reduce the reserved amount.…”
Section: Space Efficient Collectorsmentioning
confidence: 99%
“…Brain imaging big data is a collection of data that has a wide range of sources, diverse types, complex structures, and potential value, and is difficult to apply common methods of processing and analysis, after integrating its own characteristics such as regional, seasonal, diversity, and periodicity of agriculture [1]. Brain imaging big data retains the basic characteristics of big data, such as huge volume, variety, low value, fast processing speed, high veracity and high complexity, and big data application research in brain imaging is still relatively small [2].…”
Section: Introductionmentioning
confidence: 99%
“…However, the source of network public opinion information is diversified and heterogeneous. Effectively integrating such information to achieve better decision-making is a technical problem that needs to be solved urgently [2].…”
Section: Introductionmentioning
confidence: 99%