2016
DOI: 10.1109/tc.2015.2455979
|View full text |Cite
|
Sign up to set email alerts
|

Leveraging Data Deduplication to Improve the Performance of Primary Storage Systems in the Cloud

Abstract: With the explosive growth in data volume, the I/O bottleneck has become an increasingly daunting challenge for big data analytics in the Cloud. Recent studies have shown that moderate to high data redundancy clearly exists in primary storage systems in the Cloud. Our experimental studies reveal that data redundancy exhibits a much higher level of intensity on the I/O path than that on disks due to relatively high temporal access locality associated with small I/O requests to redundant data. Moreover, directly … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
7
0

Year Published

2016
2016
2024
2024

Publication Types

Select...
4
4
1

Relationship

0
9

Authors

Journals

citations
Cited by 47 publications
(7 citation statements)
references
References 25 publications
0
7
0
Order By: Relevance
“…Many of us lead often the impersonating lab tests using staying weights coming from three stockpiling structures on the Computer Research office, which includes two net laborers (web-vm), an email expert (mail) along with a store laborer [6]. The I/O follows have been assembled downstream of a functioning page free from each and every structure to get a term regarding three days.…”
Section: C) Crowd Deduplication With Cryptographymentioning
confidence: 99%
See 1 more Smart Citation
“…Many of us lead often the impersonating lab tests using staying weights coming from three stockpiling structures on the Computer Research office, which includes two net laborers (web-vm), an email expert (mail) along with a store laborer [6]. The I/O follows have been assembled downstream of a functioning page free from each and every structure to get a term regarding three days.…”
Section: C) Crowd Deduplication With Cryptographymentioning
confidence: 99%
“…Progressing International Data Corporation (IDC) examines demonstrate that lately the quantity of data has expanded by very nearly multiple times to 7ZB consistently and a more than 44overlap development to 35ZB is typical in the accompanying ten years [4]. Dealing with the data storm on storage to help (close) consistent data examination turns into an undeniably basic test for Big Data investigation in the Server, particularly for VM stages where the absolute number and strength of little documents overpower the I/O information route in the Server [6].…”
Section: Introductionmentioning
confidence: 98%
“…However, the size of the mapping table increased due to the introduction of the secondary mapping table. A. Gupta et al [13,14] used the value locality inside SSDs to reduce their write load while realizing deduplication. However, evaluating the efficiency of deduplication is impossible because they did not discuss fingerprint generation.…”
Section: State Of the Artmentioning
confidence: 99%
“…Numerous researchers depicted 'Data Deduplication' is the most efficient strategy to resolve the multiple data copies in storage [6] and [7].Data deduplication (dedup) techniques are extensively used to eliminate duplicate copy of data in cloud storage [8], [9]. Rather than only reducing the storage overheads, deduplication similarly capable to optimized the bandwidth usage [10].…”
Section: Introductionmentioning
confidence: 99%