Audiovisual Speaker Identity Verification Based on Lip Motion Features

Girija Chetty, Michael Wagner

    Research output: A Conference proceeding or a Chapter in BookConference contributionpeer-review

    17 Downloads (Pure)

    Abstract

    In this paper, we propose the fusion of audio and explicit lip motion features for speaker identity verification applications. Experimental results using GMM-based speaker models indicate that audiovisual fusion with explicit lip motion information provides significant performance improvement for verifying both the speaker identity and the liveness, due to tracking of the closely coupled acoustic labial dynamics. Experiments performed on different gender specific subsets of data from the VidTIMIT and UCBN databases under clean and noisy conditions show that the best performance of 7%– 11% EER is achieved for the speaker verification task and 4%–8% EER for the liveness verification scenario.
    Original languageEnglish
    Title of host publicationProceedings of the 8th Annual Conference of the International Speech Communication Association (Interspeech 2007)
    EditorsRonald Bock, Francesca Bonin, Nick Campbell, Ronald Poppe
    Place of PublicationGermany
    PublisherInternational Speech Communication Association
    Pages2045-2048
    Number of pages4
    ISBN (Print)9781605603162
    Publication statusPublished - 2007
    EventInterspeech 2007 - 8th Annual Conference of the International Speech Communication Association - Antwerp, Belgium
    Duration: 27 Aug 200731 Aug 2007

    Conference

    ConferenceInterspeech 2007 - 8th Annual Conference of the International Speech Communication Association
    Country/TerritoryBelgium
    CityAntwerp
    Period27/08/0731/08/07

    Fingerprint

    Dive into the research topics of 'Audiovisual Speaker Identity Verification Based on Lip Motion Features'. Together they form a unique fingerprint.

    Cite this