Proceedings of the 22nd European MPI Users' Group Meeting 2015
DOI: 10.1145/2802658.2802669
|View full text |Cite
|
Sign up to set email alerts
|

An MPI Halo-Cell Implementation for Zero-Copy Abstraction

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
6
0

Year Published

2016
2016
2024
2024

Publication Types

Select...
4
2
2

Relationship

0
8

Authors

Journals

citations
Cited by 9 publications
(6 citation statements)
references
References 13 publications
0
6
0
Order By: Relevance
“…The idea for a zero-copy framework for ghost cell exchanges has been discussed in Besnard et al (2015). Similar to our work, the authors strip away the node-local copying of ghost cell data and instead directly access the inner ghost cell data structures of node-local (and neighboring) ranks.…”
Section: A Strategy For Better Interoperability Between Gaspi and mentioning
confidence: 84%
See 2 more Smart Citations
“…The idea for a zero-copy framework for ghost cell exchanges has been discussed in Besnard et al (2015). Similar to our work, the authors strip away the node-local copying of ghost cell data and instead directly access the inner ghost cell data structures of node-local (and neighboring) ranks.…”
Section: A Strategy For Better Interoperability Between Gaspi and mentioning
confidence: 84%
“…This implementation relies on a “threads as processes” (MPC) MPI implementation, while we start from a more general approach; we directly use MPI shared memory across multiple processes and run the solver (wherever we subsequently need to exchange corresponding data) in these shared windows. Our work also conceptually extends the work in Besnard et al (2015), as we implement a model where communication (not just computation) is visible for all processes in the shared window. Furthermore, the implementation for the pipelined Allreduce, which is presented in Section 4.4, would not be feasible with the approach from Besnard et al (2015), as all processes require access to node-local communication to optimally sustain the pipeline.…”
Section: A Strategy For Better Interoperability Between Gaspi and mentioning
confidence: 91%
See 1 more Smart Citation
“…Common optimizations for improving scaling on HPC systems include options for combining MPI data exchanges for a number of arrays (e.g., vector or tensor compo-nents) or increasing the width of the grid halo region (see, e.g. [3]) for reducing latency of MPI communications. The latter allows to reduce the number of calls to MPI functions but at the cost of additional computational overhead, which may be negligible when the size of the problem on MPI-process is comparatively small.…”
Section: Methodsmentioning
confidence: 99%
“…A more intrusive approach is to get rid of halo intranode halo buffers in the users code and directly expose the neighbouring intranode processor's memory. So the intranode transfer becomes simple direct memory-to-memory copy of halo data without any intermediate buffers (zero-copy ) [5].…”
Section: Shhalo Framework For Shared-memory Halo Communicationmentioning
confidence: 99%