No Cover Image

Journal article 904 views 373 downloads

Hierarchical Brain Network for Face and Voice Integration of Emotion Expression

Jodie Davies-Thompson Orcid Logo, Giulia V Elli, Mohamed Rezk, Stefania Benetti, Markus van Ackeren, Olivier Collignon

Cerebral Cortex, Volume: 29, Issue: 9, Pages: 3590 - 3605

Swansea University Author: Jodie Davies-Thompson Orcid Logo

Check full text

DOI (Published version): 10.1093/cercor/bhy240

Abstract

The brain has separate specialized computational units to process faces and voices located in occipital and temporal cortices. However, humans seamlessly integrate signals from the faces and voices of others for optimal social interaction. How are emotional expressions, when delivered by different s...

Full description

Published in: Cerebral Cortex
ISSN: 1047-3211 1460-2199
Published: Oxford University Press (OUP) 2019
Online Access: Check full text

URI: https://cronfa.swan.ac.uk/Record/cronfa45285
Tags: Add Tag
No Tags, Be the first to tag this record!
Abstract: The brain has separate specialized computational units to process faces and voices located in occipital and temporal cortices. However, humans seamlessly integrate signals from the faces and voices of others for optimal social interaction. How are emotional expressions, when delivered by different sensory modalities (faces and voices), integrated in the brain? In this study, we characterized the brains' response to faces, voices, and combined face-voice information (congruent, incongruent), which varied in expression (neutral, fearful). Using a whole-brain approach, we found that only the right posterior superior temporal sulcus (rpSTS) responded more to bimodal stimuli than to face or voice alone but only when the stimuli contained emotional expression. Face- and voice-selective regions of interest, extracted from independent functional localizers, similarly revealed multisensory integration in the face-selective rpSTS only; further, this was the only face-selective region that also responded significantly to voices. Dynamic causal modeling revealed that the rpSTS receives unidirectional information from the face-selective fusiform face area, and voice-selective temporal voice area, with emotional expression affecting the connection strength. Our study promotes a hierarchical model of face and voice integration, with convergence in the rpSTS, and that such integration depends on the (emotional) salience of the stimuli.
Keywords: emotional expression, faces, fMRI, multisensory, voice
College: Faculty of Medicine, Health and Life Sciences
Issue: 9
Start Page: 3590
End Page: 3605