Audiovisual Speaker Identity Verification Based on Lip Motion Features

Girija Chetty, Michael Wagner

Research output: A Conference proceeding or a Chapter in BookConference contributionpeer-review

35 Downloads (Pure)

Abstract

In this paper, we propose the fusion of audio and explicit lip motion features for speaker identity verification applications. Experimental results using GMM-based speaker models indicate that audiovisual fusion with explicit lip motion information provides significant performance improvement for verifying both the speaker identity and the liveness, due to tracking of the closely coupled acoustic labial dynamics. Experiments performed on different gender specific subsets of data from the VidTIMIT and UCBN databases under clean and noisy conditions show that the best performance of 7%– 11% EER is achieved for the speaker verification task and 4%–8% EER for the liveness verification scenario.
Original languageEnglish
Title of host publicationProceedings of the 8th Annual Conference of the International Speech Communication Association (Interspeech 2007)
EditorsRonald Bock, Francesca Bonin, Nick Campbell, Ronald Poppe
Place of PublicationGermany
PublisherInternational Speech Communication Association
Pages2045-2048
Number of pages4
ISBN (Print)9781605603162
Publication statusPublished - 2007
EventInterspeech 2007 - 8th Annual Conference of the International Speech Communication Association - Antwerp, Belgium
Duration: 27 Aug 200731 Aug 2007

Conference

ConferenceInterspeech 2007 - 8th Annual Conference of the International Speech Communication Association
Country/TerritoryBelgium
CityAntwerp
Period27/08/0731/08/07

Fingerprint

Dive into the research topics of 'Audiovisual Speaker Identity Verification Based on Lip Motion Features'. Together they form a unique fingerprint.

Cite this