2020
DOI: 10.14778/3384345.3384358
|View full text |Cite
|
Sign up to set email alerts
|

Traversing large graphs on GPUs with unified memory

Abstract: Due to the limited capacity of GPU memory, the majority of prior work on graph applications on GPUs has been restricted to graphs of modest sizes that fit in memory. Recent hardware and software advances make it possible to address much larger host memory transparently as a part of a feature known as unified virtual memory. While accessing host memory over an interconnect is understandably slower, the problem space has not been sufficiently explored in the context of a challenging workload with low computation… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
8
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
3

Relationship

0
8

Authors

Journals

citations
Cited by 43 publications
(8 citation statements)
references
References 46 publications
0
8
0
Order By: Relevance
“…Evaluation of Azimov's algorithm shows that it is possible to improve performance by using GPGPU because operations of linear algebra can be efficiently implemented on GPGPU (Mishin et al, 2019;Terekhov et al, 2020). Moreover, for practical reasons, it is interesting to provide a multi-GPU version of the algorithm and to utilize unified memory, which is suitable for linear algebra based processing of out-of-GPGPU-memory data and traversing on large graphs (Chien et al, 2019;Gera et al, 2020).…”
Section: Discussionmentioning
confidence: 99%
“…Evaluation of Azimov's algorithm shows that it is possible to improve performance by using GPGPU because operations of linear algebra can be efficiently implemented on GPGPU (Mishin et al, 2019;Terekhov et al, 2020). Moreover, for practical reasons, it is interesting to provide a multi-GPU version of the algorithm and to utilize unified memory, which is suitable for linear algebra based processing of out-of-GPGPU-memory data and traversing on large graphs (Chien et al, 2019;Gera et al, 2020).…”
Section: Discussionmentioning
confidence: 99%
“…We evaluate competing systems' performance on three applications, BC, LL, and NCP. To keep consistent with previous work [1,23,53], we configure the three applications as follows.…”
Section: Methodsmentioning
confidence: 99%
“…The algorithm samples the starting nodes from the graph. Therefore, in our evaluation, we randomly sample a batch of 100 source vertices for each graph [23].…”
Section: Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…Third, an extremely large graph, which drives the needs of graph sampling and random walk, usually goes beyond the size of GPU memory. While there exists an array of solutions for GPU-based large graph processing, namely, unified memory [26], topology-aware partition [27] and vertex-range based partitions [28], graph sampling and random walk algorithms, which require all the neighbors of a vertex to present in order to compute the selection probability, exhibit stringent requirement on the partitioning methods. In the meantime, the asynchronous and out-of-order nature of graph sampling and random walk provides some unique optimization opportunities for out-of-memory sampling, which are neither shared nor explored by traditional out-of-memory systems.…”
Section: Introductionmentioning
confidence: 99%