Title:
|
Audio, video and multimodal person identification in a smart room
|
Author:
|
Luque, Jordi; Morros, R.; Garde, I.; Anguita, Jan; Farrús, Mireia; Macho, D.; Marqués López, Fernando; Martínez, C.; Vilaplana, V.; Hernando, Javier
|
Abstract:
|
Comunicació presentada a: The First International Evaluation Workshop on Classification of Events, Activities and Relationships, CLEAR 2006, celebrat a Southampton, Regne Unit, el 6 i 7 d'abril de 2006. |
Abstract:
|
In this paper, we address the modality integration issue on the example of a smart room environment aiming at enabling person identification by combining speech and 2D face images. First we introduce the monomodal audio and video identification techniques and then we present the use of combined input speech and face images for person identification. The various sensory modalities, speech and faces, are processed both individually and jointly. It’s shown that the multimodal approach results in improved performance in the identification of the participants. |
Abstract:
|
This work has been partially sponsored by the EC-funded project CHIL (IST-2002-506909) and by the Spanish Government-funded project ACESCA (TIN2005-08852). |
Subject(s):
|
-Multimodality -Speaker recognition |
Rights:
|
© Springer The final publication is available at Springer via http://dx.doi.org/10.1007/978-3-540-69568-4_23 |
Document type:
|
Conference Object Article - Accepted version |
Published by:
|
Springer
|
Share:
|
|