Research in photonic computing has flourished due to the proliferation of optoelectronic components on photonic integration platforms. Photonic integrated circuits have enabled ultrafast artificial neural networks, providing a framework for a new class of information processing machines. Algorithms running on such hardware have the potential to address the growing demand for machine learning and artificial intelligence, in areas such as medical diagnosis, telecommunications, and high-performance and scientific computing. In parallel, the development of neuromorphic electronics has highlighted challenges in that domain, in particular, related to processor latency. Neuromorphic photonics offers subnanosecond latencies, providing a complementary opportunity to extend the domain of artificial intelligence. Here, we review recent advances in integrated photonic neuromorphic systems, discuss current and future challenges, and outline the advances in science and technology needed to meet those challenges.Conventional computers are organized around a centralized processing architecture (i.e. with a central processor and memory), which is suited to run sequential, digital, procedure-based programs. Such an architecture is inefficient for computational models that are distributed, massively parallel, and adaptive, most notably, those used for neural networks in artificial intelligence (AI). AI is an attempt to approach human level accuracy on these tasks that are challenging for traditional computers but easy for humans. Major achievements have been realized by machine learning (ML) algorithms based on neural networks [1], which process information in a distributed fashion and adapt to past inputs rather than being explicitly designed by a programmer. ML has had an impact on many aspects of our lives with applications ranging from translating languages [2] to cancer diagnosis [3]. Neuromorphic engineering is partly an attempt to move elements of ML and AI algorithms to hardware that reflects their massively distributed nature. Matching hardware to algorithms leads potentially to faster and more energy efficient information processing. Neuromorphic hardware is also applied to problems outside of ML, such as robot control, mathematical programming, and neuroscientific hypothesis testing [4,5]. Massively distributed hardware relies heavily-more so than other computer architectures-on massively parallel interconnections between lumped elements (i.e. neurons). Dedicated metal wiring for every connection is not practical. Therefore, current state-of-the-art neuromorphic electronics use some form of shared digital communication bus that is timedivision multiplexed, trading bandwidth for interconnectivity [4]. Optical interconnects could negate this trade-off and thus have the potential to accelerate ML and neuromorphic computing.Light is established as the communication medium of telecom and datacenters, but it has not yet found widespread use in information processing and computing. The same properties that allow optoelectronic ...