2001
DOI: 10.1016/s0167-8191(01)00073-4
|View full text |Cite
|
Sign up to set email alerts
|

Towards a fast parallel sparse symmetric matrix–vector multiplication

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
34
0
3

Year Published

2005
2005
2021
2021

Publication Types

Select...
5
2
2

Relationship

0
9

Authors

Journals

citations
Cited by 38 publications
(37 citation statements)
references
References 5 publications
0
34
0
3
Order By: Relevance
“…Bell and Garland consider several methods, including a variation of ELLPACK that differs from ours [3]. They split the storage between an ELLPACK and coordinate format to reduce its footprint, a novel variant of other previously proposed splitting methods [9,13,20]. At the same time, Baskaran and Bordawekar proposed a general compile-and run-time infrastructure, evaluated for SpMV [2].…”
Section: Related Researchmentioning
confidence: 99%
“…Bell and Garland consider several methods, including a variation of ELLPACK that differs from ours [3]. They split the storage between an ELLPACK and coordinate format to reduce its footprint, a novel variant of other previously proposed splitting methods [9,13,20]. At the same time, Baskaran and Bordawekar proposed a general compile-and run-time infrastructure, evaluated for SpMV [2].…”
Section: Related Researchmentioning
confidence: 99%
“…Furthermore, storage of the entire H is not necessary, significantly reducing storage requirements. The sparsity and highly structured form of A matrices would suggest the use of sparse iterative techniques, yet research has shown the A matrix to be a worst case scenario in many respects, that elicits poor performance from all standard iterative methods [26]- [28]. To this end, the algorithms developed in ScalIT address the shortcomings, and have proven to be effective and highly parallelizable.…”
Section: The "A Matrix" Form and Scalit Methodologymentioning
confidence: 99%
“…Other tuning techniques include diagonal cache blocking [30], the detection of diagonal substructures [11], the exploitation of symmetries [21], and optimizations for specific higher-level kernels, such as sparse triangular solve [34]. The impact of prefetching on SMVM performance was previously explored in [31] and [37].…”
Section: Related Workmentioning
confidence: 99%