Multimodal Fusion for Robust Identity Authentication: Role of Liveness Checks

Girija Chetty, Emdad Hossain

Research output: A Conference proceeding or a Chapter in BookChapter

Abstract

Most of the current biometric identity authentication systems currently deployed are based on modeling the identity of a person based on unimodal information, i.e. face, voice, or fingerprint features. Also, many current interactive civilian remote human computer interaction applications are based on speech based voice features, which achieve significantly lower performance for operating environments with low signal-to-noise ratios (SNR). For a long time, use of acoustic information alone has been a great success for several automatic speech processing applications such as automatic speech transcription or speaker authentication, while face identification systems based visual information alone from faces also proved to be of equally successful. However, in adverse operating environments, performance of either of these systems could be suboptimal. Use of both visual and audio information can lead to better robustness, as they can provide complementary secondary clues that can help in the analysis of the primary biometric signals (Potamianos et al (2004)). The joint analysis of acoustic and visual speech can improve the robustness of automatic speech recognition systems (Liu et al (2002), Gurbuz et al (2002)
Original languageEnglish
Title of host publicationAdvanced Biometric Technologies
EditorsGirija Chetty, Jucheng Yang
Place of PublicationCroatia
PublisherIn-Tech
Pages3-20
Number of pages18
Edition1
ISBN (Print)9789533074870
Publication statusPublished - 2011

Fingerprint Dive into the research topics of 'Multimodal Fusion for Robust Identity Authentication: Role of Liveness Checks'. Together they form a unique fingerprint.

  • Cite this

    Chetty, G., & Hossain, E. (2011). Multimodal Fusion for Robust Identity Authentication: Role of Liveness Checks. In G. Chetty, & J. Yang (Eds.), Advanced Biometric Technologies (1 ed., pp. 3-20). In-Tech.