2015
DOI: 10.1145/2686875
|View full text |Cite
|
Sign up to set email alerts
|

GP-SIMD Processing-in-Memory

Abstract: AMIR MORAD, LEONID YAVITS, and RAN GINOSAR, Technion GP-SIMD, a novel hybrid general-purpose SIMD computer architecture, resolves the issue of data synchronization by in-memory computing through combining data storage and massively parallel processing. GP-SIMD employs a two-dimensional access memory with modified SRAM storage cells and a bit-serial processing unit per each memory row. An analytic performance model of the GP-SIMD architecture is presented, comparing it to associative processor and to convention… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
17
0

Year Published

2016
2016
2022
2022

Publication Types

Select...
4
3
3

Relationship

0
10

Authors

Journals

citations
Cited by 37 publications
(17 citation statements)
references
References 56 publications
0
17
0
Order By: Relevance
“…NMC is enabled by new memory technologies, such as 3D-stacked memories [5,62,65,66,81], and also by cache-coherent interconnects [24,88,93], which allow close integration of processing units and memory units. Depending on the applications and systems of interest (e.g., [13,14,15,22,23,27,29,31,40,42,49,50,53,58,61,68,69,71,74,77,87]), prior works propose di erent types of near-memory processing units, such as general-purpose CPU cores [13,16,27,28,29,39,64,69,78,83,86], GPU cores [44,52,80,106], recon gurable units [41,55,57,90], or xed-function units…”
Section: Related Workmentioning
confidence: 99%
“…NMC is enabled by new memory technologies, such as 3D-stacked memories [5,62,65,66,81], and also by cache-coherent interconnects [24,88,93], which allow close integration of processing units and memory units. Depending on the applications and systems of interest (e.g., [13,14,15,22,23,27,29,31,40,42,49,50,53,58,61,68,69,71,74,77,87]), prior works propose di erent types of near-memory processing units, such as general-purpose CPU cores [13,16,27,28,29,39,64,69,78,83,86], GPU cores [44,52,80,106], recon gurable units [41,55,57,90], or xed-function units…”
Section: Related Workmentioning
confidence: 99%
“…The PIM core could be an SIMD machine, a GPU-like multithreading machine [4], a reconfigurable array, many core systems, etc. References [26,27] also states that an SIMD/VLIW/vector processor is fit for the data-processing unit in a PIM system. The host CPU and PIM cores share the same physical memory.…”
Section: Target Architecturementioning
confidence: 99%
“…To work around the limitations presented by cache coherence, most prior works assume a limited amount of sharing between the PIM kernels and the processor threads of an application. Thus, they sidestep coherence by employing solutions that restrict PIM to execute on non-cacheable data (e.g., [2,47,52,149,243]) or force processor cores to flush or not access any data that could potentially be used by PIM (e.g., [3,4,28,47,51,59,67,68,163,172,195,196,199,200]). In fact, the IMPICA accelerator design, described in Section 3, falls into the latter category.…”
Section: Lazypim: An Efficient Cache Coherence Mechanism For Processi...mentioning
confidence: 99%