2022 International Electron Devices Meeting (IEDM) 2022
DOI: 10.1109/iedm45625.2022.10019569
|View full text |Cite
|
Sign up to set email alerts
|

Deep learning acceleration in 14nm CMOS compatible ReRAM array: device, material and algorithm co-optimization

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
14
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
5
1

Relationship

0
6

Authors

Journals

citations
Cited by 12 publications
(14 citation statements)
references
References 12 publications
0
14
0
Order By: Relevance
“…This enables the training of DNNs on extremely noisy devices and reduces the number of states required for the auxiliary device to barely 10 levels. Experimental demonstrations of the algorithms on ECRAM and ReRAM arrays have been reported, which verify the successful training with practical devices. These studies suggest that algorithmic remedies can overcome nonidealities in memory devices and improve software-level classification accuracy (Figure ).…”
Section: Algorithm and Architecturementioning
confidence: 59%
“…This enables the training of DNNs on extremely noisy devices and reduces the number of states required for the auxiliary device to barely 10 levels. Experimental demonstrations of the algorithms on ECRAM and ReRAM arrays have been reported, which verify the successful training with practical devices. These studies suggest that algorithmic remedies can overcome nonidealities in memory devices and improve software-level classification accuracy (Figure ).…”
Section: Algorithm and Architecturementioning
confidence: 59%
“…In our recent work, we also showed that the Tiki-Taka algorithm improves the training accuracy compared to the SGD method. We could perform the training of a fully connected network (FCN) on a reduced Modified National Institute of Standards and Technology (MNIST) data set, but devices with enhanced properties are required for scaling to larger networks.…”
mentioning
confidence: 88%
“…In previous works, the weight update was performed row by row 14 or device by device. 15 In our recent work, 16 NN training with parallel crossbar updates was demonstrated. We implemented artificial synapses using resistive random access memory (RRAM) devices.…”
mentioning
confidence: 99%
See 1 more Smart Citation
“…It was shown (Gokmen, 2021) that the filtering stage greatly lowers the device requirements, in particular the effective number of device states: only 15 states are enough to train an LSTM to acceptable accuracy. This relaxed device requirements now make it possible to think about current ReRAM devices for in-memory training (Gong et al, 2022). This Tiki-Taka version 2 (TTv2) algorithm is our baseline comparison 3 .…”
Section: Ttv2 Algorithmmentioning
confidence: 99%