2017 IEEE International Conference on Cluster Computing (CLUSTER) 2017
DOI: 10.1109/cluster.2017.80
|View full text |Cite
|
Sign up to set email alerts
|

TAPIOCA: An I/O Library for Optimized Topology-Aware Data Aggregation on Large-Scale Supercomputers

Abstract: Reading and writing data efficiently from storage system is necessary for most scientific simulations to achieve good performance at scale. Many software solutions have been developed to decrease the I/O bottleneck. One well-known strategy, in the context of collective I/O operations, is the twophase I/O scheme. This strategy consists of selecting a subset of processes to aggregate contiguous pieces of data before performing reads/writes. In this paper, we present TAPIOCA, an MPI-based library implementing an … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
10
0

Year Published

2018
2018
2024
2024

Publication Types

Select...
4
3
1

Relationship

1
7

Authors

Journals

citations
Cited by 25 publications
(11 citation statements)
references
References 17 publications
0
10
0
Order By: Relevance
“…With an increasing number of memory storage layers and complexity of storage system interactions, the issue of I/O performance is getting imperative and significantly hinders the overall performance of applications . Research efforts have been made to relax the POSIX semantics and alleviate the I/O bottleneck from high‐level libraries (eg, HDF5, netCDF, ADIOS), I/O middleware (eg, MPI‐IO, TAPIOCA), to I/O forwarding layers . All of these provide an array‐based data model to organize the data and define data access semantics.…”
Section: Related Workmentioning
confidence: 99%
“…With an increasing number of memory storage layers and complexity of storage system interactions, the issue of I/O performance is getting imperative and significantly hinders the overall performance of applications . Research efforts have been made to relax the POSIX semantics and alleviate the I/O bottleneck from high‐level libraries (eg, HDF5, netCDF, ADIOS), I/O middleware (eg, MPI‐IO, TAPIOCA), to I/O forwarding layers . All of these provide an array‐based data model to organize the data and define data access semantics.…”
Section: Related Workmentioning
confidence: 99%
“…Other researchers took the routing mechanism of BG/Q into consideration when issued sparse data access [12]. Tessier et al [13] used a different approach which combines an optimized buffering system and a topology-aware aggregators mapping algorithm targeting any kind of architecture, so the new algorithm can be easily extended. Tsujita et al [14] introduced a topology-aware data aggregation scheme which takes care of processes rank layout across compute nodes and rearranges the data collection sequence during the shuffle phase in order to mitigate network contention.…”
Section: Related Workmentioning
confidence: 99%
“…The conventional parallel I/O stack consists of high-level libraries (HDF5 [23], NetCDF [24], ADIOS [25], etc. ), I/O middleware (MPI-IO [26], TAPIOCA [27]), and I/O forwarding layer [28]. Several research efforts have focused on relaxing the POSIX semantics and on defining new data models in these layers.…”
Section: Related Workmentioning
confidence: 99%