Proceedings of the Twenty-Fifth Annual ACM Symposium on Parallelism in Algorithms and Architectures 2013
DOI: 10.1145/2486159.2486196
|View full text |Cite
|
Sign up to set email alerts
|

Communication optimal parallel multiplication of sparse random matrices

Abstract: Parallel algorithms for sparse matrix-matrix multiplication typically spend most of their time on inter-processor communication rather than on computation, and hardware trends predict the relative cost of communication will only increase. Thus, sparse matrix multiplication algorithms must minimize communication costs in order to scale to large processor counts.In this paper, we consider multiplying sparse matrices corresponding to Erdős-Rényi random graphs on distributed-memory parallel machines. We prove a ne… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

2
40
0

Year Published

2013
2013
2023
2023

Publication Types

Select...
4
3
2

Relationship

1
8

Authors

Journals

citations
Cited by 57 publications
(42 citation statements)
references
References 32 publications
2
40
0
Order By: Relevance
“…Hence, the FLOP-to-communication ratio decreases as 1/ √ P in these weak scaling experiments. This observation can also be shown theoretically [22] The bottleneck is more pronounced on networks where the effective bandwidth decreases with increasing P . Similar observations hold for indexing overhead, which also becomes more important as the number of multiplication steps increases.…”
Section: Performance Limitssupporting
confidence: 71%
“…Hence, the FLOP-to-communication ratio decreases as 1/ √ P in these weak scaling experiments. This observation can also be shown theoretically [22] The bottleneck is more pronounced on networks where the effective bandwidth decreases with increasing P . Similar observations hold for indexing overhead, which also becomes more important as the number of multiplication steps increases.…”
Section: Performance Limitssupporting
confidence: 71%
“…T AB: Optimizing sparse matrixmatrix multiplication is an active area of research [17], [18]; state-of-the-art implementations are bound by the memory bandwidth and heavily underutilize the compute resources.…”
Section: ) Optimizing Res = Amentioning
confidence: 99%
“…Parallelisation and indexing techniques for sparse matrices multiplication were implemented by B u l u c and G i l b e r t [7]. The communication overhead problem of sparse matrices multiplication was solved by B a l l a r d et al [8]. The parallelisation technique for sparse tensor matrix multiplication was proposed by S m i t h et al [9].…”
Section: Introductionmentioning
confidence: 99%
“…The parallelisation technique for sparse tensor matrix multiplication was proposed by S m i t h et al [9]. The above approaches [7][8][9] are not suitable for Big Data applications. Proper care should be taken by the programmer regarding the data distribution, replication, load balancing, communication overhead etc.…”
Section: Introductionmentioning
confidence: 99%