Dense crossbar arrays of non-volatile memory (NVM) devices represent one possible path for implementing massively-parallel and highly energy-efficient neuromorphic computing systems. We first review recent advances in the application of NVM devices to three computing paradigms: spiking neural networks (SNNs), deep neural networks (DNNs), and 'Memcomputing'. In SNNs, NVM synaptic connections are updated by a local learning rule such as spike-timing-dependent-plasticity, a computational approach directly inspired by biology. For DNNs, NVM arrays can represent matrices of synaptic weights, implementing the matrix-vector multiplication needed for algorithms such as backpropagation in an analog yet massively-parallel fashion. This approach could provide significant improvements in power and speed compared to GPU-based DNN training, for applications of commercial significance. We then survey recent research in which different types of NVM devices-including phase change memory, conductive-bridging RAM, filamentary and nonfilamentary RRAM, and other NVMs-have been proposed, either as a synapse or as a neuron, for use within a neuromorphic computing application. The relevant virtues and limitations of these devices are assessed, in terms of properties such as conductance dynamic range, (non)linearity and (a)symmetry of conductance response, retention, endurance, required switching power, and device variability.
Neural-network training can be slow and energy intensive, owing to the need to transfer the weight data for the network between conventional digital memory chips and processor chips. Analogue non-volatile memory can accelerate the neural-network training algorithm known as backpropagation by performing parallelized multiply-accumulate operations in the analogue domain at the location of the weight data. However, the classification accuracies of such in situ training using non-volatile-memory hardware have generally been less than those of software-based training, owing to insufficient dynamic range and excessive weight-update asymmetry. Here we demonstrate mixed hardware-software neural-network implementations that involve up to 204,900 synapses and that combine long-term storage in phase-change memory, near-linear updates of volatile capacitors and weight-data transfer with 'polarity inversion' to cancel out inherent device-to-device variations. We achieve generalization accuracies (on previously unseen data) equivalent to those of software-based training on various commonly used machine-learning test datasets (MNIST, MNIST-backrand, CIFAR-10 and CIFAR-100). The computational energy efficiency of 28,065 billion operations per second per watt and throughput per area of 3.6 trillion operations per second per square millimetre that we calculate for our implementation exceed those of today's graphical processing units by two orders of magnitude. This work provides a path towards hardware accelerators that are both fast and energy efficient, particularly on fully connected neural-network layers.
We survey the current state of phase change memory (PCM), a non-volatile
solid-state memory technology built around the large electrical contrast
between the highly-resistive amorphous and highly-conductive crystalline states
in so-called phase change materials. PCM technology has made rapid progress in
a short time, having passed older technologies in terms of both sophisticated
demonstrations of scaling to small device dimensions, as well as integrated
large-array demonstrators with impressive retention, endurance, performance and
yield characteristics.
We introduce the physics behind PCM technology, assess how its
characteristics match up with various potential applications across the
memory-storage hierarchy, and discuss its strengths including scalability and
rapid switching speed. We then address challenges for the technology, including
the design of PCM cells for low RESET current, the need to control
device-to-device variability, and undesirable changes in the phase change
material that can be induced by the fabrication procedure. We then turn to
issues related to operation of PCM devices, including retention,
device-to-device thermal crosstalk, endurance, and bias-polarity effects.
Several factors that can be expected to enhance PCM in the future are
addressed, including Multi-Level Cell technology for PCM (which offers higher
density through the use of intermediate resistance states), the role of coding,
and possible routes to an ultra-high density PCM technology.Comment: Review articl
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.