2011
DOI: 10.1007/978-3-642-23397-5_41
|View full text |Cite
|
Sign up to set email alerts
|

Iterative Sparse Matrix-Vector Multiplication for Integer Factorization on GPUs

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
9
0

Year Published

2012
2012
2019
2019

Publication Types

Select...
3
2
2

Relationship

1
6

Authors

Journals

citations
Cited by 7 publications
(9 citation statements)
references
References 13 publications
0
9
0
Order By: Relevance
“…As reported in [17], experimental results on GeForce GTX 480 show that SpMV kernel with the cache blocking method is 5x faster than the unblocked CSR kernel in the best case. And in [18], we can see speedups between 4 and 8 on a single GPU for a number of tested NFS matrices compared to an optimized multi-core implementation. We can strongly believe that our algorithm is at least exceeding the conventional CPU algorithms.…”
Section: Cost Analysismentioning
confidence: 87%
See 2 more Smart Citations
“…As reported in [17], experimental results on GeForce GTX 480 show that SpMV kernel with the cache blocking method is 5x faster than the unblocked CSR kernel in the best case. And in [18], we can see speedups between 4 and 8 on a single GPU for a number of tested NFS matrices compared to an optimized multi-core implementation. We can strongly believe that our algorithm is at least exceeding the conventional CPU algorithms.…”
Section: Cost Analysismentioning
confidence: 87%
“…For the densest part, we use the dense format CSR and SLE for the sparse part to improve the performance. More details can be found in Schmidt's paper [18].…”
Section: B Spare Matrix Formatsmentioning
confidence: 98%
See 1 more Smart Citation
“…These works have explored implementing efficiently SpMV over real numbers. Schmidt et al [19] proposed an optimized matrix format to accelerate exact SpMV over GF (2), that can be used in the linear algebra step of the Number Field Sieve (NFS) for integer factorization [22]. Boyer et al [8] have adapted SpMV kernels over small finite fields and rings Z/mZ, where they used double-precision floating-point numbers to represent ring elements.…”
Section: Sparse-matrix-vector Product On Gpusmentioning
confidence: 99%
“…Its parallel implementation using the processing power of multi-core CPUs or a single-node GPU too are not too good to be used even for medium-sized matrices of few hundred thousand rows and columns. Matrix data to be processed, received from fields such as climatology, seismology, cryptography [4] and many other fields, range in sizes of multi-million rows and columns and often require real-time processing.…”
Section: Introductionmentioning
confidence: 99%