2015
DOI: 10.1109/tpds.2014.2370055
|View full text |Cite
|
Sign up to set email alerts
|

Sparse Matrix Multiplication On An Associative Processor

Abstract: Abstract-Sparse matrix multiplication is an important component of linear algebra computations. Implementing sparse matrix multiplication on an associative processor (AP) enables high level of parallelism, where a row of one matrix is multiplied in parallel with the entire second matrix, and where the execution time of vector dot product does not depend on the vector size. Four sparse matrix multiplication algorithms are explored in this paper, combining AP and baseline CPU processing to various levels. They a… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
9
0

Year Published

2016
2016
2023
2023

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 29 publications
(9 citation statements)
references
References 38 publications
(76 reference statements)
0
9
0
Order By: Relevance
“…The scaling of CMOS associative processors [16], [17] is limited due to the CMOS CAM density. However, an MRAM AP cell is at least an order-of-magnitude smaller, thus paving the way for associative in-memory computing at scale.…”
Section: B Associative Processormentioning
confidence: 99%
See 1 more Smart Citation
“…The scaling of CMOS associative processors [16], [17] is limited due to the CMOS CAM density. However, an MRAM AP cell is at least an order-of-magnitude smaller, thus paving the way for associative in-memory computing at scale.…”
Section: B Associative Processormentioning
confidence: 99%
“…Since modern datasets become increasingly sparse, the ability of computers to properly process sparse data (for example, not wasting time and energy on fetching and multiplying zero-data elements) becomes a critical requirement. An AP holds a significant intrinsic advantage in sparse data processing: sparse data in one of the compressed formats (such as compressed sparse row or compressed sparse column) can be processed almost as efficiently as dense data [17].…”
Section: Application Spacementioning
confidence: 99%
“…APs have been explored for many applications such as matrix multiplication [11], [12], fast Fourier transform (FFT) [13], discrete Fourier transform (DCT) and video application [43], DNA sequence alignment [14], stencil applications [44], convolution operation [9], solution of path problems [45], optimum branchings [46], databases applications [47] and computer vision [48]. The old applications need to be revisited and re-evaluate under the new AP design approaches and technologies besides exploring new applications that could benefit from the AP.…”
Section: B Promising Applicationsmentioning
confidence: 99%
“…For this reason, their best usage is in applications that have an inherent SIMD (single-instruction multiple-data) computational pattern. Fast Fourier transform (FFT) [15], DNA sequence alignment [16], stencil [17], and matrix multiplication [18], [19] are some example applications that can benefit from AP. In an analogy, APs can be considered as a next step on the path of the CPU (central processing unit) to GPU (graphical processing unit) transformation.…”
Section: Content Addressable Memorymentioning
confidence: 99%