SC18: International Conference for High Performance Computing, Networking, Storage and Analysis 2018
DOI: 10.1109/sc.2018.00018
|View full text |Cite
|
Sign up to set email alerts
|

Distributed-Memory Hierarchical Compression of Dense SPD Matrices

Abstract: We present a distributed memory algorithm for the hierarchical compression of symmetric positive definite (SPD) matrices. Our method is based on GOFMM, an algorithm that appeared in doi:10.1145/3126908.3126921. For many SPD matrices, GOFMM enables compression and approximate matrix-vector multiplication that for many matrices can reach N log N time-as opposed to N 2 required for a dense matrix. But GOFMM supports only shared memory parallelism. In this paper, we use the message passing interface (MPI) and exte… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

0
11
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
3
2
2

Relationship

1
6

Authors

Journals

citations
Cited by 14 publications
(11 citation statements)
references
References 21 publications
0
11
0
Order By: Relevance
“…Strong scaling: In Figure 2 (#1, #2, #3, #4, #5, #6, #7, #8, #9, #10, #11, and #12), we use a 2 24 -by-2 24 Gaussian kernel matrix generated with a synthetic 6-D point dataset. Using this matrix, we perform strong scaling experiments using up to 6,144 Skylake cores (128 compute nodes, using one MPI process per node and 48 OpenMP threads).…”
Section: Resultsmentioning
confidence: 99%
See 3 more Smart Citations
“…Strong scaling: In Figure 2 (#1, #2, #3, #4, #5, #6, #7, #8, #9, #10, #11, and #12), we use a 2 24 -by-2 24 Gaussian kernel matrix generated with a synthetic 6-D point dataset. Using this matrix, we perform strong scaling experiments using up to 6,144 Skylake cores (128 compute nodes, using one MPI process per node and 48 OpenMP threads).…”
Section: Resultsmentioning
confidence: 99%
“…Our method is based on GOFMM [23], [24], which constructs a hierarchical matrix factorization for K, using only entries in K. (We review GOFMM in §II). In particular, GOFMM constructs a matrix K (hereby "compresses") using O(N log N ) entries from K such that K − K ≤ K (for a user-defined tolerance > 0) and a matvec with K requires as low as O(N ) work.…”
Section: Introductionmentioning
confidence: 99%
See 2 more Smart Citations
“…Xia [31] used a similar strategy for compressing blocks in a nested dissection solver, and the STRUMPACK package [27,14] also uses this randomized sampling strategy for compression of fronts into HSS form. Shared and distributed memory hierarchical compression of SPD matrices into a weak admissibility format are described in [33,34].…”
mentioning
confidence: 99%