2006
DOI: 10.1007/11758501_30
|View full text |Cite
|
Sign up to set email alerts
|

Performance Improvement of Sparse Matrix Vector Product on Vector Machines

Abstract: Abstract. Many applications based on finite element and finite difference methods include the solution of large sparse linear systems using preconditioned iterative methods. Matrix vector multiplication is one of the key operations that has a significant impact on the performance of any iterative solver. In this paper, recent developments in sparse storage formats on vector machines are reviewed. Then, several improvements to memory access in the sparse matrix vector product are suggested. Particularly, algori… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2007
2007
2008
2008

Publication Types

Select...
2

Relationship

1
1

Authors

Journals

citations
Cited by 2 publications
(2 citation statements)
references
References 5 publications
0
2
0
Order By: Relevance
“…Using vector registers to reduce memory operations for loading and storing the result vector further improves the performance of JAD based sparse MVP to 2.2 GFlop/s. Further optimizations result in a maximum performance of 20% vector peak (which is 16 GFlop/s) for sparse MVP on NEC SX-8 [6].…”
Section: Average Vector Lengthmentioning
confidence: 99%
See 1 more Smart Citation
“…Using vector registers to reduce memory operations for loading and storing the result vector further improves the performance of JAD based sparse MVP to 2.2 GFlop/s. Further optimizations result in a maximum performance of 20% vector peak (which is 16 GFlop/s) for sparse MVP on NEC SX-8 [6].…”
Section: Average Vector Lengthmentioning
confidence: 99%
“…Thus, small blocks can be formed by grouping the equations at each grid point. Operating on such dense blocks considerably reduces the amount of indirect addressing required for sparse MVP [6]. This improves the performance of the kernel dramatically on vector machines [9] and also remarkably on superscalar architectures [10,11].…”
Section: Block-based Linear Iterative Solver (Blis)mentioning
confidence: 99%