2020 Design, Automation &Amp; Test in Europe Conference &Amp; Exhibition (DATE) 2020
DOI: 10.23919/date48585.2020.9116464
|View full text |Cite
|
Sign up to set email alerts
|

TDO-CIM: Transparent Detection and Offloading for Computation In-memory

Abstract: Computation in-memory is a promising non-von Neumann approach aiming at completely diminishing the data transfer to and from the memory subsystem. Although a lot of architectures have been proposed, compiler support for such architectures is still lagging behind. In this paper, we close this gap by proposing an end-to-end compilation flow for in-memory computing based on the LLVM compiler infrastructure. Starting from sequential code, our approach automatically detects, optimizes, and offloads kernels suitable… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
4
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
1
1

Relationship

0
7

Authors

Journals

citations
Cited by 9 publications
(4 citation statements)
references
References 26 publications
(25 reference statements)
0
4
0
Order By: Relevance
“…Unfortunately, previous CIM-oriented compilation frameworks overlook the aforementioned deployment issues. For example, frameworks like TC-CIM [28], TDO-CIM [29], and OCC [30] regard the CIM macro as an abstract model similar to the tensor core, which is a type of processing element of graphic processing units (GPUs) [31]. However, operator incompatibility and macro utilization are not evaluated and discussed in these works.…”
Section: (C) Most Previous Work Evaluate Cnns On Lightweightmentioning
confidence: 99%
See 2 more Smart Citations
“…Unfortunately, previous CIM-oriented compilation frameworks overlook the aforementioned deployment issues. For example, frameworks like TC-CIM [28], TDO-CIM [29], and OCC [30] regard the CIM macro as an abstract model similar to the tensor core, which is a type of processing element of graphic processing units (GPUs) [31]. However, operator incompatibility and macro utilization are not evaluated and discussed in these works.…”
Section: (C) Most Previous Work Evaluate Cnns On Lightweightmentioning
confidence: 99%
“…In this section, the deployment results of convolution-based algorithms are analyzed. The evaluation is conducted based on a cost model of an SRAM-CIM system [28][29][30]. First, the proposed weight mapping strategies are evaluated across different convolution layers.…”
Section: Experimental Evaluationmentioning
confidence: 99%
See 1 more Smart Citation
“…There have been attempts at making everything transparent to the programmer by reworking existing parallel computing frameworks such as Apache Spark, MapReduce, [51][52][53] and MPI [10]. Alternately, compilers automatically identify and transform these regions appropriately to a form suitable for the targeted PIM architecture [28,31,64]. Our work falls into the last camp.…”
Section: Programming Models For Pimmentioning
confidence: 99%