2022
DOI: 10.48550/arxiv.2207.07886
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

An Experimental Evaluation of Machine Learning Training on a Real Processing-in-Memory System

Abstract: Training machine learning algorithms is a computationally intensive process, which is frequently memory-bound due to repeatedly accessing large training datasets. As a result, processor-centric systems (e.g., CPU, GPU) suffer from costly data movement between memory units and processing units, which consumes large amounts of energy and execution cycles. Memory-centric computing systems, i.e., computing systems with processing-in-memory (PIM) capabilities, can alleviate this data movement bottleneck.Our goal is… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
2
1

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(2 citation statements)
references
References 98 publications
0
2
0
Order By: Relevance
“…Compared to inference, training is by far more compute and memory intensive and therefore the major challenge to address in consumer devices. The training procedure of Neural Networks (NNs), in fact, relies on BP [21] which is memory-bounded [40]. This is due to the storage of NN's activations, resulting in slow and energy-consuming operations.…”
Section: On-device Learningmentioning
confidence: 99%
“…Compared to inference, training is by far more compute and memory intensive and therefore the major challenge to address in consumer devices. The training procedure of Neural Networks (NNs), in fact, relies on BP [21] which is memory-bounded [40]. This is due to the storage of NN's activations, resulting in slow and energy-consuming operations.…”
Section: On-device Learningmentioning
confidence: 99%
“…Many previous works investigate how to provide new functionality using compute-capable memories based on conventional (e.g., [1,2,36,40,43,72,99,108,110]) and emerging memory technologies (e.g., [9,27,56,59,68,73,101,111,116,117,119,121,139,144]) to help solve the data movement overheads in today's systems. These works propose new functionality in at least three major categories: (1) support for logical operations (e.g., [26,73,86,110,119,121,139,144]), (2) support for complex operations, functions, and applications (e.g., [1,36,72,89,111,112,116,117,143]), and (3) programming and system support for the integration and adoption of such accelerators (e.g., [2,3,9,19,27,55,…”
Section: Computation-in-memory Acceleratorsmentioning
confidence: 99%