Emotion communication must be robust to interference from a noisy environment. One safeguard against interference is crossmodal redundancy-for example, conveying the same information using both sound and movement. Emotion perceivers should therefore be adapted to e iciently detect crossmodal correspondences, increasing the likelihood that emotion signals will be understood. One possible such adaptation is the use of a single neural code for both auditory and visual information. To investigate this, we tested two hypotheses: ( ) that distinct auditory and visual brain areas represent emotion expressions using the same parameters, and ( ) that auditory and visual expressions of emotion are represented together in one brain area using a supramodal neural code. We presented emotion expressions during functional magnetic resonance imaging (N= , scan hrs/participant) and tested these hypotheses using representational similarity analysis (Kriegeskorte & Kievit, ). A single model of stimulus features and emotion content fit brain activity in both auditory and visual areas, supporting hypothesis ( ), and posterior superior temporal gyrus represented both auditory and visual emotion expressions, supporting hypothesis ( ). These results hold for both discrete and mixed (e.g., Happy-Sad) emotional expressions. Surprisingly, further exploratory analysis showed auditory and visual areas represent stimulus features and emotion content even when stimuli are presented in each area's non-preferred modality.