Proceedings of the 52nd Annual IEEE/ACM International Symposium on Microarchitecture 2019
DOI: 10.1145/3352460.3358291
|View full text |Cite
|
Sign up to set email alerts
|

SparTen

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
6
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 191 publications
(24 citation statements)
references
References 22 publications
0
6
0
Order By: Relevance
“…through pruning can help mitigate the computation complexity [26][27][28][29][30][31]. More efficient ways of using the computation hardware infrastructure, including static or dynamic exploitation of value sparsity to avoid unnecessary computation, are also being developed [32][33][34][35][36][37][38]. These are all valuable efforts to mitigate the complexity explosion and to help sustain the continuation of the current productive trends.…”
Section: Figure Taken Frommentioning
confidence: 99%
“…through pruning can help mitigate the computation complexity [26][27][28][29][30][31]. More efficient ways of using the computation hardware infrastructure, including static or dynamic exploitation of value sparsity to avoid unnecessary computation, are also being developed [32][33][34][35][36][37][38]. These are all valuable efforts to mitigate the complexity explosion and to help sustain the continuation of the current productive trends.…”
Section: Figure Taken Frommentioning
confidence: 99%
“…The values are delivered to an array of multipliers, and the resulting scattered products are summed using a dedicated interconnection mesh. Sparten [70] is based on SCNN architecture, but it improves the distribution of the operations to the multipliers to reduce the overhead. EIE [71] compresses the weights with the CSC scheme and has zero-skipping ability for null activations.…”
Section: Accelerators With Sparsity Exploitationmentioning
confidence: 99%
“…While we explore further workloads in Section 3.3, our main motivation is accelerating general sparse LA on a set of established tensor formats. Unlike approaches focused on low sparsities [10], [11], [17], SSSRs target flexibility and scalability to achieve notable speedups across a wider sparsity range (see Section 4.1). SSSRs support iterating along the major axis of any tensor format where said axis is given by two arrays: a value array storing nonzero values and an index array storing their positions.…”
Section: Accelerable Formats and Operationsmentioning
confidence: 99%
“…Most general-purpose instruction set architecture (ISA) extensions consider only one-sided sparsity [8], [9], and those targeting sparse-sparse workloads incur large area impacts and cannot efficiently handle the former [6]. Sparse ML solutions often rely on custom, domain-specific formats and low or structured sparsity [10], [11], while more general sparse LA accelerators [12], [13] incur significant silicon area impacts compared to their host systems and other solutions.…”
Section: Introductionmentioning
confidence: 99%