2021 IEEE International Solid- State Circuits Conference (ISSCC) 2021
DOI: 10.1109/isscc42613.2021.9365989
|View full text |Cite
|
Sign up to set email alerts
|

15.4 A 5.99-to-691.1TOPS/W Tensor-Train In-Memory-Computing Processor Using Bit-Level-Sparsity-Based Optimization and Variable-Precision Quantization

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
6
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
4
1
1

Relationship

0
6

Authors

Journals

citations
Cited by 28 publications
(6 citation statements)
references
References 4 publications
0
6
0
Order By: Relevance
“…CMOS ANN hardware for inference has already achieved high energy efficiency of more than 600 TOPS/W with MAC operations based on a CIM architecture [29] by limiting the bit-width of the synaptic weights as shown in Table 4. The energy efficiency is higher than that of ANN hardware using ReRAM [10], [28].…”
Section: Discussionmentioning
confidence: 99%
“…CMOS ANN hardware for inference has already achieved high energy efficiency of more than 600 TOPS/W with MAC operations based on a CIM architecture [29] by limiting the bit-width of the synaptic weights as shown in Table 4. The energy efficiency is higher than that of ANN hardware using ReRAM [10], [28].…”
Section: Discussionmentioning
confidence: 99%
“…If a deep learning accelerator runs a big model with limited on-chip memory resources, TTD is a good choice to help it avoid a large size of memory footprint. However, it may be necessary to design a domain-specific accelerator for handling the increased computation cost and improving energy efficiency, by taking advantage of reduced external memory accesses, such as [38], [39].…”
Section: Train-ttd-trainmentioning
confidence: 99%
“…To utilize both digital and CIM architecture for optimal performance on the various layer characteristics, a low-power SoC, comprising a digital neural network accelerator, an analog CIM core and a RISC-V CPU, is designed to realize simultaneous execution of subsequent layers and high/low precision on corresponding cores [25] . Algorithm-hardware co-design method is utilized for higher energy efficiency.…”
Section: System-level Cim Chipsmentioning
confidence: 99%