Prosody for the eyes: quantifying visual prosody using guided principal component analysis

Erin Cvejic, Jeesun Kim, Chris Davis, Guillaume Gibert

Research output: Chapter in Book / Conference PaperConference Paperpeer-review

19 Citations (Scopus)

Abstract

Although typically studied as an auditory phenomenon, prosody can also be conveyed by the visual speech signal, through increased movements of articulators during speech production, or through eyebrow and rigid head movements. This paper aimed to quantify such visual correlates of prosody. Specifically, the study was concerned with measuring the visual correlates of prosodic focus and prosodic phrasing. In the experiment, four participants' speech and face movements were recorded while they completed a dialog exchange task with an interlocutor. Acoustic analysis showed that prosodic contrasts differed on duration, pitch and intensity parameters, which is consistent with previous findings in the literature. The visual data was processed using guided principal component analysis. The results showed that compared to the broad focused statement condition, speakers produced greater movement on both articulatory and non-articulatory parameters for prosodically focused and intonated words.
Original languageEnglish
Title of host publicationProceedings of the 11th Annual Conference of the International Speech Communication Association, INTERSPEECH 2010: Makuhari, Japan, 26-30 September 2010
PublisherInternational Speech Communication Association
Pages1433-1436
Number of pages4
Publication statusPublished - 2010
EventInternational Speech Communication Association. Conference -
Duration: 9 Sept 2012 → …

Conference

ConferenceInternational Speech Communication Association. Conference
Period9/09/12 → …

Fingerprint

Dive into the research topics of 'Prosody for the eyes: quantifying visual prosody using guided principal component analysis'. Together they form a unique fingerprint.

Cite this