Abstract
The experiment investigated how the addition of emotion information from the voice affects the identification of facial emotion. We presented whole face, upper face, and lower face displays and examined correct recognition rates and patterns of response confusions for auditory-visual (AV), auditory-only (AO), and visual-only (VO) expressive speech. Emotion recognition accuracy was superior for AV compared to unimodal presentation. The pattern of response confusions differed across the unimodal conditions and across display type. For AV presentation, a response confusion only occurred when such a confusion was present in each modality separately, thus response confusions were reduced compared to unimodal presentations. Emotion space (calculated from the confusion data) differed across display types for the VO presentations but was more similar for the AV ones indicating that the addition of the auditory information acted to harmonize the various VO response patterns. These results are discussed with respect to how bimodal emotion recognition combines auditory and visual information.
Original language | English |
---|---|
Pages (from-to) | 902-921 |
Number of pages | 20 |
Journal | Visual Cognition |
Volume | 20 |
Issue number | 8 |
DOIs | |
Publication status | Published - 2012 |
Keywords
- bimodal processing
- acoustic impedance
- speech
- hearing
- emotion processing
- face expression
- voice expression
- perception
- visual stimulation
- visual information
- stimulus response
- vision