2017 IEEE High Performance Extreme Computing Conference (HPEC) 2017
DOI: 10.1109/hpec.2017.8091048
|View full text |Cite
|
Sign up to set email alerts
|

Design and implementation of parallel PageRank on multicore platforms

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
7
0

Year Published

2017
2017
2024
2024

Publication Types

Select...
7
1

Relationship

4
4

Authors

Journals

citations
Cited by 18 publications
(7 citation statements)
references
References 11 publications
0
7
0
Order By: Relevance
“…Leveraging sparse linear algebra for graph processing is the focus of the GraphBLAS project, which aims at defining operations on graphs through the language of linear algebra [11], and it offers early implementations for both CPU and GPU [4,22]. Highly tuned implementations of PPR exploit the graph data-layout to maximize cache usage [25], or employ multi-machine setups to process trillions of edges [26]. Green-Marl [8] and GraphIt [24] implements PPR using Domain-Specific Languages (DSLs) that abstract the intricacies of graph processing, and optimized to fully exploits the CPU hardware.…”
Section: Cpu and Gpu Implementationsmentioning
confidence: 99%
“…Leveraging sparse linear algebra for graph processing is the focus of the GraphBLAS project, which aims at defining operations on graphs through the language of linear algebra [11], and it offers early implementations for both CPU and GPU [4,22]. Highly tuned implementations of PPR exploit the graph data-layout to maximize cache usage [25], or employ multi-machine setups to process trillions of edges [26]. Green-Marl [8] and GraphIt [24] implements PPR using Domain-Specific Languages (DSLs) that abstract the intricacies of graph processing, and optimized to fully exploits the CPU hardware.…”
Section: Cpu and Gpu Implementationsmentioning
confidence: 99%
“…1 for our analysis. Our reasons are two fold: 1) the two key types of memory accesses which we define below are also found in other optimized algorithms and 2) several of the optimized algorithms, such as [26], [4], have been designed to reduce the number of random memory accesses, which makes it harder to stress and evaluate the memory system with this type of algorithm.…”
Section: A Experiments Setupmentioning
confidence: 99%
“…Algorithmic optimizations have been developed to improve the spatial locality of graph analytics kernels by reducing the number of cache misses [4], [26], [25], [6], but these approaches are typically application-dependent.…”
Section: Related Workmentioning
confidence: 99%
“…Binning can be used in conjunction with both Vertex-centric or Edge-centric paradigms. Zhou et al [43,44] use a custom sorted edge list with Edgecentric processing to reduce DRAM row activations and improve memory performance. However, their sorting mechanism introduces a non-trivial pre-processing cost and imposes the use of COO format.…”
Section: Related Workmentioning
confidence: 99%