2021
DOI: 10.1109/jxcdc.2021.3057856
|View full text |Cite
|
Sign up to set email alerts
|

Ferroelectric Field-Effect Transistor-Based 3-D NAND Architecture for Energy-Efficient on-Chip Training Accelerator

Abstract: Different from the deep neural network (DNN) inference process, the training process produces huge amount of intermediate data to compute the new weights of the network. Generally, on-chip global buffer (e.g. SRAM cache) has limited capacity because of its low memory density, therefore the off-chip DRAM access is inevitable during the training sequences. In this work, a novel ferroelectric field effect transistor (FeFET) based 3D NAND architecture for on-chip training accelerator is proposed. The reduced perip… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2

Citation Types

0
3
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5

Relationship

1
4

Authors

Journals

citations
Cited by 11 publications
(3 citation statements)
references
References 32 publications
0
3
0
Order By: Relevance
“…However, extending the CTFbased design for intensive write training would be challenging due to the high programming voltage (>20 V), the slow write latency (>10 µs) and limited endurance characteristics (<10 5 cycles) of the FN-tunneling-based program and erase scheme. Consequently, the FeFET-based 3D NAND architecture was regarded as an attractive means to overcome these limitations for in situ training [74]. Owing to its ultra-high density, a 3D NAND-based design for energy-efficient DNN training was feasible with minimized DRAM access by fully storing the huge amount of intermediate data on-chip.…”
Section: D Nand-based Fefet For Trainingmentioning
confidence: 99%
See 1 more Smart Citation
“…However, extending the CTFbased design for intensive write training would be challenging due to the high programming voltage (>20 V), the slow write latency (>10 µs) and limited endurance characteristics (<10 5 cycles) of the FN-tunneling-based program and erase scheme. Consequently, the FeFET-based 3D NAND architecture was regarded as an attractive means to overcome these limitations for in situ training [74]. Owing to its ultra-high density, a 3D NAND-based design for energy-efficient DNN training was feasible with minimized DRAM access by fully storing the huge amount of intermediate data on-chip.…”
Section: D Nand-based Fefet For Trainingmentioning
confidence: 99%
“…For the VMM operations within the FeFET-based 3D NAND, a BL input scheme is used (explained in section 4.2). The details of the data flow and VMM operations of each training process (feed forward, error calculation, weight gradient calculation, and weight update) in each WL group are discussed in more detail in [74]. Subsequently, evaluation of the energy efficiency was performed for each on-chip training step using the DNN + NeuroSim framework [22], as shown in figure 17.…”
Section: D Nand-based Fefet For Trainingmentioning
confidence: 99%
“…Thus, the required number of memory arrays and chip size should be further increased for the implementation of complex NN models which contain multiple layers. One of the solutions for this issue can be the use of three-dimensional (3D) memory structures, which can stack the memory elements without increasing the area of the chip 17,18 . Alternatively, there are approaches using conventional memory devices such as NOR flash, NAND flash, and AND flash to implement NNs [19][20][21][22] .…”
mentioning
confidence: 99%