Deep convolutional neural networks (CNNs) resemble the hierarchically organised neural representations in the primate visual ventral stream. However, these models typically disregard the temporal dynamics experimentally observed in these areas. For instance, alpha oscillations dominate the dynamics of the human visual cortex, yet the computational relevance of oscillations is rarely considered in artificial neural networks (ANNs). We propose an ANN that embraces oscillatory dynamics with the computational purpose of converting simultaneous inputs, presented at two different locations, into a temporal code. The network was trained to classify three individually presented letters. Post-training, we added semi-realistic temporal dynamics to the hidden layer, introducing relaxation dynamics in the hidden units as well as pulsed inhibition mimicking neuronal alpha oscillations. Without these dynamics, the trained network correctly classified individual letters but produced a mixed output when presented with two letters simultaneously, elucidating a bottleneck problem. When introducing refraction and oscillatory inhibition, the output nodes corresponding to the two stimuli activated sequentially, ordered along the phase of the inhibitory oscillations. Our model provides a novel approach for implementing multiplexing in ANNs. It further produces experimentally testable predictions of how the primate visual system handles competing stimuli.