Proceedings of the 10th Parallel Data Storage Workshop 2015
DOI: 10.1145/2834976.2834985
|View full text |Cite
|
Sign up to set email alerts
|

Comparative I/O workload characterization of two leadership class storage clusters

Abstract: The Oak Ridge Leadership Computing Facility (OLCF) is a leader in large-scale parallel file system development, design, deployment and continuous operation. For the last decade, the OLCF has designed and deployed two large center-wide parallel file systems. The first instantiation, Spider 1, served the Jaguar supercomputer and its predecessor, Spider 2, now serves the Titan supercomputer, among many other OLCF computational resources. The OLCF has been rigorously collecting file and storage system statistics f… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
6
0

Year Published

2016
2016
2023
2023

Publication Types

Select...
3
2
2

Relationship

1
6

Authors

Journals

citations
Cited by 28 publications
(6 citation statements)
references
References 12 publications
0
6
0
Order By: Relevance
“…The read size is small compared other studies with 90% reads smaller than 1KB. In other studies 90% of reads are smaller than 1KB in web caches [2], 100KB in HPC [3], 1MB in HPC [13], 4MB in consumer cloud [20], 10MB in video delivery [25] and 15MB in HBase [14]. Figure 7 depicts that distributions of the two traces W1 and W2 are not too different.…”
Section: Read Sizesmentioning
confidence: 83%
See 1 more Smart Citation
“…The read size is small compared other studies with 90% reads smaller than 1KB. In other studies 90% of reads are smaller than 1KB in web caches [2], 100KB in HPC [3], 1MB in HPC [13], 4MB in consumer cloud [20], 10MB in video delivery [25] and 15MB in HBase [14]. Figure 7 depicts that distributions of the two traces W1 and W2 are not too different.…”
Section: Read Sizesmentioning
confidence: 83%
“…We have collected from Databricks a combined Spark and S3 workload trace spanning over 6 months of operation, and designed a method to process and analyze this trace. Our method focuses HPC ✓ ✓ Abad 2012 [1] MapReduce + HDFS ✓ ✓ ✓ ✓ ✓ Chen 2012 [5] MapReduce + HDFS ✓ ✓ ✓ Atikoglu 2012 [2] Web Cache ✓ ✓ ✓ ✓ Liu 2013 [20] Consumer Cloud ✓ ✓ ✓ Harter 2014 [14] Messaging + HDFS ✓ ✓ ✓ Gunasekaran 2015 [13] HPC ✓ ✓ Summers 2016 [25] Video Delivery ✓ ✓…”
Section: Resultsmentioning
confidence: 99%
“…The behavior of I/O depending on the transfer size is important in order to estimate I/O performance in scientific applications in HPC. According to the study [4] of the storage clusters Spider and Spider II at the Oak Ridge Leadership Facility (OLCF), about 50% of the requested transfer sizes are below 16KB and 50% between 512KB and 1MB in size. In figure 5 we can see that for 1MB transfer size, both local storage layers already have reached a region of maximal performance.…”
Section: Discussionmentioning
confidence: 99%
“…The flash cells within SSDs can be rewritten a finite number of times before they are no longer able to reliably store data, and as a result, SSDs are only warranted for a finite number of drive writes per day (DWPD) over their service life. Since HPC file systems have historically been subject to write-intensive workloads [10,22], the endurance requirements of SSDs in HPC environments have been a cause of concern. To date, most large-scale flash deployments in HPC have resorted to using extreme-endurance SSDs (5-10 DWPD for a fiveyear period) to ensure that the SSDs do not fail before the end of the overall system's service life [11,19].…”
Section: Drive Endurancementioning
confidence: 99%