Prosody for the eyes: quantifying visual prosody using guided principal component analysis

Erin Cvejic, Jeesun Kim, Chris Davis, Guillaume Gibert

    Research output: Chapter in Book / Conference PaperConference Paperpeer-review

    19 Citations (Scopus)

    Abstract

    Although typically studied as an auditory phenomenon, prosody can also be conveyed by the visual speech signal, through increased movements of articulators during speech production, or through eyebrow and rigid head movements. This paper aimed to quantify such visual correlates of prosody. Specifically, the study was concerned with measuring the visual correlates of prosodic focus and prosodic phrasing. In the experiment, four participants' speech and face movements were recorded while they completed a dialog exchange task with an interlocutor. Acoustic analysis showed that prosodic contrasts differed on duration, pitch and intensity parameters, which is consistent with previous findings in the literature. The visual data was processed using guided principal component analysis. The results showed that compared to the broad focused statement condition, speakers produced greater movement on both articulatory and non-articulatory parameters for prosodically focused and intonated words.
    Original languageEnglish
    Title of host publicationProceedings of the 11th Annual Conference of the International Speech Communication Association, INTERSPEECH 2010: Makuhari, Japan, 26-30 September 2010
    PublisherInternational Speech Communication Association
    Pages1433-1436
    Number of pages4
    Publication statusPublished - 2010
    EventInternational Speech Communication Association. Conference -
    Duration: 9 Sept 2012 → …

    Conference

    ConferenceInternational Speech Communication Association. Conference
    Period9/09/12 → …

    Fingerprint

    Dive into the research topics of 'Prosody for the eyes: quantifying visual prosody using guided principal component analysis'. Together they form a unique fingerprint.

    Cite this