Most of the current biometric identity authentication systems currently deployed are based on modeling the identity of a person based on unimodal information, i.e. face, voice, or fingerprint features. Also, many current interactive civilian remote human computer interaction applications are based on speech based voice features, which achieve significantly lower performance for operating environments with low signal-to-noise ratios (SNR). For a long time, use of acoustic information alone has been a great success for several automatic speech processing applications such as automatic speech transcription or speaker authentication, while face identification systems based visual information alone from faces also proved to be of equally successful. However, in adverse operating environments, performance of either of these systems could be suboptimal. Use of both visual and audio information can lead to better robustness, as they can provide complementary secondary clues that can help in the analysis of the primary biometric signals (Potamianos et al (2004)). The joint analysis of acoustic and visual speech can improve the robustness of automatic speech recognition systems (Liu et al (2002), Gurbuz et al (2002)
|Title of host publication||Advanced Biometric Technologies|
|Editors||Girija Chetty, Jucheng Yang|
|Place of Publication||Croatia|
|Number of pages||18|
|Publication status||Published - 2011|
Chetty, G., & Hossain, E. (2011). Multimodal Fusion for Robust Identity Authentication: Role of Liveness Checks. In G. Chetty, & J. Yang (Eds.), Advanced Biometric Technologies (1 ed., pp. 3-20). In-Tech.