2022
DOI: 10.1109/mm.2021.3096236
|View full text |Cite
|
Sign up to set email alerts
|

ECIM: Exponent Computing in Memory for an Energy-Efficient Heterogeneous Floating-Point DNN Training Processor

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
4
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
4
2

Relationship

0
6

Authors

Journals

citations
Cited by 10 publications
(4 citation statements)
references
References 6 publications
0
4
0
Order By: Relevance
“…Yet most existing SRAM-based IMC structures have low data density 8 . The data density in [41,42] is lower than 0.3Mb mm −2 . This inevitably causes more off-chip memory communication during training, and can hardly scale out for large-scale training tasks at low cost.…”
Section: Introductionmentioning
confidence: 79%
See 1 more Smart Citation
“…Yet most existing SRAM-based IMC structures have low data density 8 . The data density in [41,42] is lower than 0.3Mb mm −2 . This inevitably causes more off-chip memory communication during training, and can hardly scale out for large-scale training tasks at low cost.…”
Section: Introductionmentioning
confidence: 79%
“…Recently, there are several works employing floating-point formats for IMC designs: Tu et al [41] developed a digital processor handling GeMM between integers and FP numbers, reaching 14 TFLOPS/W efficiency computing GeMM with BFloat16 format. Lee et al [42] designed an SRAM-based IMC circuit processing GeMM between BFloat16, which separates exponent and fraction storage with the best computing efficiency at 1.43 TFLOPS W −1 ; Lee et al [43] give a DRAM-based near-memory computing design with high throughput at 1 TFLOPS per chip. Nevertheless, efficient in-memory simultaneous processing of mantissa and exponents are still not explored.…”
Section: Introductionmentioning
confidence: 99%
“…The architecture uses bfloat16DNN to train the processor to optimize the exponential computation and to improve energy efficiency and reduce memory consumption. The resulting computational architecture achieves an energy efficiency of 13.7 TFLOPS/W when processing data [6]. Wang C et al developed a DNN-based optimized read voltage value strategy for expressing the relationship between voltage distribution and read voltage threshold.…”
Section: Related Workmentioning
confidence: 99%
“…In equation (3.12), p is the quantity of neurons in the hidden layer, n is the quantity of neurons in the input layer, and a is the mediation parameter, which takes values in the range of [1][2][3][4][5][6][7][8][9][10]. To sum up, this research first carries out the assessment index confirmation of the TSC risk of enterprises, then data and normalization of the index, then constructs a multi-level DNN model to calculate the characteristics of the processed index, and finally judges the TSC risk degree of the enterprise according to the calculation results.…”
Section: Construction Of a Multi-level Dnn-based Risk Prediction Mode...mentioning
confidence: 99%