Proceedings of the 20th ACM SIGPLAN/SIGBED International Conference on Languages, Compilers, and Tools for Embedded Systems 2019
DOI: 10.1145/3316482.3326351
|View full text |Cite
|
Sign up to set email alerts
|

Optimizing tensor contractions for embedded devices with racetrack memory scratch-pads

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
9
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
5
1

Relationship

3
3

Authors

Journals

citations
Cited by 11 publications
(9 citation statements)
references
References 49 publications
0
9
0
Order By: Relevance
“…The work in [138] investigates the layouts of highdimensional data structures such as tensors in RTM-based SPMs. For the tensor contraction operation, an optimized data layout reduces the number of shifts by 50% compared to a naïve layout.…”
Section: B Software Techniques For Minimizing Shiftmentioning
confidence: 99%
See 1 more Smart Citation
“…The work in [138] investigates the layouts of highdimensional data structures such as tensors in RTM-based SPMs. For the tensor contraction operation, an optimized data layout reduces the number of shifts by 50% compared to a naïve layout.…”
Section: B Software Techniques For Minimizing Shiftmentioning
confidence: 99%
“…For instance, different RTM topologies and structures differ in their error patterns which need to be analyzed at the architectural level. Similarly, at the compiler level, the memory access patterns of applications can be reordered from higher compiler abstractions, e.g., from a polyhedral model or by additional semantic information from domain-specific languages [138]. There is a need to investigate a runtime system that is flexible to adapt to various flavors of the racetrack (single DW versus multiple DWs; horizontal versus vertical racetrack) memories and different application characteristics (latency versus bandwidthsensitive applications).…”
Section: Hw/sw Codesignmentioning
confidence: 99%
“…If the local adjacency of a vertex with the left group is greater than that of the right group, then it is added to left group (cf. lines [12][13][14]. Otherwise, the vertex is added to the right group (cf.…”
Section: The Shiftsreduce Heuristicmentioning
confidence: 99%
“…The overall performance and energy benefits of (Chen, ShiftsReduce, and IGA-Ours) compared to OFU translate to (22.2%, 25.4%, and 31.7%) and (12.4%, 17.5%, and 26.4%), respectively. The suitability of RMs compared to other memory technologies such as SRAM, STT-MRAM, and DRAM has already been established [13,30,48].…”
Section: Summary Performance and Energy Analysismentioning
confidence: 99%
See 1 more Smart Citation