INTAIRACT: Joint hand gesture and fingertip classification for touchless interaction

dc.contributor
Universitat Politècnica de Catalunya. Departament de Teoria del Senyal i Comunicacions
dc.contributor
Universitat Politècnica de Catalunya. GPI - Grup de Processament d'Imatge i Vídeo
dc.contributor.author
Suau Cuadros, Xavier
dc.contributor.author
Alcoverro Vidal, Marcel
dc.contributor.author
López Méndez, Adolfo
dc.contributor.author
Ruiz Hidalgo, Javier
dc.contributor.author
Casas Pla, Josep Ramon
dc.date.issued
2012
dc.identifier
Suau, X. [et al.]. INTAIRACT: Joint hand gesture and fingertip classification for touchless interaction. A: European Conference on Computer Vision. "Computer Vision – ECCV 2012. Workshops and Demonstrations". Florència: Springer, 2012, p. 602-606.
dc.identifier
978-3-642-33884-7
dc.identifier
https://hdl.handle.net/2117/18278
dc.identifier
10.1007/978-3-642-33885-4_62
dc.description.abstract
In this demo we present intAIRact, an online hand-based touchless interaction system. Interactions are based on easy-to-learn hand gestures, that combined with translations and rotations render a user friendly and highly configurable system. The main advantage with respect to existing approaches is that we are able to robustly locate and identify fingertips. Hence, we are able to employ a simple but powerful alphabet of gestures not only by determining the number of visible fingers in a gesture, but also which fingers are being observed. To achieve such a system we propose a novel method that jointly infers hand gestures and fingertip locations using a single depth image from a consumer depth camera. Our approach is based on a novel descriptor for depth data, the Oriented Radial Distribution (ORD) [1]. On the one hand, we exploit the ORD for robust classification of hand gestures by means of efficient k-NN retrieval. On the other hand, maxima of the ORD are used to perform structured inference of fingertip locations. The proposed method outperforms other state-of-the-art approaches both in gesture recognition and fingertip localization. An implementation of the ORD extraction on a GPU yields a real-time demo running at approximately 17fps on a single laptop
dc.description.abstract
Peer Reviewed
dc.description.abstract
Postprint (published version)
dc.format
5 p.
dc.format
application/pdf
dc.language
eng
dc.publisher
Springer
dc.relation
http://link.springer.com/chapter/10.1007/978-3-642-33885-4_62
dc.rights
http://creativecommons.org/licenses/by-nc-nd/3.0/es/
dc.rights
Restricted access - publisher's policy
dc.rights
Attribution-NonCommercial-NoDerivs 3.0 Spain
dc.subject
Àrees temàtiques de la UPC::Informàtica::Sistemes d'informació::Interacció home-màquina
dc.subject
Àrees temàtiques de la UPC::Enginyeria de la telecomunicació
dc.subject
Human-computer interaction
dc.subject
Interacció persona-ordinador
dc.title
INTAIRACT: Joint hand gesture and fingertip classification for touchless interaction
dc.type
Conference lecture


Fitxers en aquest element

FitxersGrandàriaFormatVisualització

No hi ha fitxers associats a aquest element.

Aquest element apareix en la col·lecció o col·leccions següent(s)

E-prints [72954]