2012 SC Companion: High Performance Computing, Networking Storage and Analysis 2012
DOI: 10.1109/sc.companion.2012.372
|View full text |Cite
|
Sign up to set email alerts
|

A Case for Scaling HPC Metadata Performance through De-specialization

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
5
0

Year Published

2015
2015
2020
2020

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 10 publications
(5 citation statements)
references
References 6 publications
0
5
0
Order By: Relevance
“…This is caused by complex distributed locking of central data structures (generally managed by a metadata server instance) that are required to be accessed in parallel [31]. The file system community presented various techniques for handling metadata [12,32,35,36,37,38], but this challenge is still prevailing and is becoming an even bigger challenge for upcoming data-science applications.…”
Section: Related Workmentioning
confidence: 99%
“…This is caused by complex distributed locking of central data structures (generally managed by a metadata server instance) that are required to be accessed in parallel [31]. The file system community presented various techniques for handling metadata [12,32,35,36,37,38], but this challenge is still prevailing and is becoming an even bigger challenge for upcoming data-science applications.…”
Section: Related Workmentioning
confidence: 99%
“…Typically, general-purpose PFSs distribute data across all available storage targets. As this technique works well for data, it does not achieve the same throughput when handling metadata [5], [28], although the file system community presented various techniques to tackle this challenge [3], [13], [25], [26], [41], [42]. The performance limitation can be attributed to the sequentialization enforced by underlying POSIX semantics which is particularly degrading throughput when a huge number of files is created in a single directory from multiple processes.…”
Section: Related Workmentioning
confidence: 99%
“…FPFS also implements management for huge directories, or hugedirs for short. These directories are common for some HPC applications, such as those that create a file per thread/process, and those that use a directory as a light‐weight database (eg, check pointing) . To manage hugedirs, FPFS proposes a dynamic distribution of their entries among multiple OSD+s .…”
Section: Overview Of Fpfsmentioning
confidence: 99%