2010 10th IEEE/ACM International Conference on Cluster, Cloud and Grid Computing 2010
DOI: 10.1109/ccgrid.2010.81
|View full text |Cite
|
Sign up to set email alerts
|

From Sparse Matrix to Optimal GPU CUDA Sparse Matrix Vector Product Implementation

Abstract: The CUDA model for GPUs presents the programmer with a plethora of different programming options. These includes different memory types, different memory access methods, and different data types. Identifying which options to use and when is a non-trivial exercise. This paper explores the effect of these different options on the performance of a routine that evaluates sparse matrix vector products. A process for analysing performance and selecting the subset of implementations that perform best is proposed. The… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
4
0

Year Published

2011
2011
2023
2023

Publication Types

Select...
3
3
1

Relationship

0
7

Authors

Journals

citations
Cited by 10 publications
(5 citation statements)
references
References 9 publications
1
4
0
Order By: Relevance
“…Our work produces results comparable to other modeling efforts of the SpMV on GPUs [6], [8], [9]. Our proposed model offers an alternative method in the prediction of SpMV execution time by using the number of memory accesses.…”
Section: Resultssupporting
confidence: 73%
See 1 more Smart Citation
“…Our work produces results comparable to other modeling efforts of the SpMV on GPUs [6], [8], [9]. Our proposed model offers an alternative method in the prediction of SpMV execution time by using the number of memory accesses.…”
Section: Resultssupporting
confidence: 73%
“…El Zein and Rendell [8] attempted to identify the best CUDA implementation using the CSR format with an experimental approach. Their technique creates multiple kernel implementations, all using different combinations of memory locations for storing the matrix.…”
Section: Literature Reviewmentioning
confidence: 99%
“…Many libraries are available, such as [21,[29][30][31][32][33][34]. Some libraries are so specialized that they handle only a single aspect of linear algebra, as seen in [35][36][37][38][39][40][41]. Each of these libraries offers different implementations.…”
Section: Inner Corementioning
confidence: 99%
“…Zein and Rendell [26] has explored the effect of these different options on the performance of a routine that evaluated sparse matrix vector products. They have proposed a process for analysing performance and selecting the subset of implementations that perform best.…”
Section: Many-core Graphics Processing Unitmentioning
confidence: 99%