Line drawings convey meaning with just a few strokes. Despite strong simplifications, humans can recognize objects depicted in such abstracted images without effort. To what degree do deep convolutional neural networks (CNNs) mirror this human ability to generalize to abstracted object images? While CNNs trained on natural images have been shown to exhibit poor classification performance on drawings, other work has demonstrated highly similar latent representations in the networks for abstracted and natural images. Here, we address these seemingly conflicting findings by analyzing the activation patterns of a CNN trained on natural images across a set of photographs, drawings, and sketches of the same objects and comparing them to human behavior. We find a highly similar representational structure across levels of visual abstraction in early and intermediate layers of the network. This similarity, however, does not translate to later stages in the network, resulting in low classification performance for drawings and sketches. We identified that texture bias in CNNs contributes to the dissimilar representational structure in late layers and the poor performance on drawings. Finally, by fine-tuning late network layers with object drawings, we show that performance can be largely restored, demonstrating the general utility of features learned on natural images in early and intermediate layers for the recognition of drawings. In conclusion, generalization to abstracted images, such as drawings, seems to be an emergent property of CNNs trained on natural images, which is, however, suppressed by domain-related biases that arise during later processing stages in the network.
Predicting actions from non-verbal cues and using them to optimise one's response behaviour (i.e. interpersonal predictive coding) is essential in everyday social interactions. We aimed to investigate the neural correlates of different cognitive processes evolving over time during interpersonal predictive coding. Thirty-nine participants watched two agents depicted by moving point-light stimuli while an electroencephalogram (EEG) was recorded. One well-recognizable agent performed either a 'communicative' or an 'individual' action. The second agent either was blended into a cluster of noise dots (i.e. present) or was entirely replaced by noise dots (i.e. absent), which participants had to differentiate. EEG amplitude and coherence analyses for theta, alpha and beta frequency bands revealed a dynamic pattern unfolding over time: Watching communicative actions was associated with enhanced coupling within medial anterior regions involved in social and mentalising processes and with dorsolateral prefrontal activation indicating a higher
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.