Full metadata record
DC poleHodnotaJazyk
dc.contributor.authorKurmi, Vinod K
dc.contributor.authorJain, Garima
dc.contributor.authorVenkatesh, KS
dc.contributor.editorSkala, Václav
dc.date.accessioned2018-05-17T08:23:14Z-
dc.date.available2018-05-17T08:23:14Z-
dc.date.issued2015
dc.identifier.citationWSCG '2015: short communications proceedings: The 23rd International Conference in Central Europe on Computer Graphics, Visualization and Computer Vision 2015 in co-operation with EUROGRAPHICS: University of West Bohemia, Plzen, Czech Republic8-12 June 2015, p. 159-166.en
dc.identifier.isbn978-80-86943-66-4
dc.identifier.issn2464-4617
dc.identifier.uriwscg.zcu.cz/WSCG2015/CSRN-2502.pdf
dc.identifier.urihttp://hdl.handle.net/11025/29678
dc.description.abstractIn this paper, we use the output of a 3D sensor (ex. Kinect from Microsoft) to capture depth images of humans making a set of predefined hand gestures in various body poses. Conventional approaches using Kinect data have been constrained by the limitation of the human detector middleware that requires close conformity to a standard near erect, legs apart, hands apart pose for the subject. Our approach also permits clutter and possible motion in the scene background, and to a limited extent, in the foreground as well. We make an important point in this work to emphasize that the recognition performance is considerably improved by a choice of hand gestures that accommodate the sensor’s specific limitations. These sensor limitations include low resolution in x and y as well as z. Hand gestures have been chosen(designed) for easy detection by seeking to detect a fingers apart, fingertip constellation with minimum computation. without, however compromising on issues of utility or ergonomy. It is shown that these gestures can be recognised in real time irrespective of visible band illumination levels, background motion, foreground clutter, user body pose, gesturing speeds and user distance. The last is of course limited by the sensor’s own range limitations. Our main contributions are the selection and design of gestures suitable for limited range, limited resolution 3D sensors and the novel method of depth slicing used to extract hand features from the background. This obviates the need for preliminary human detection and enables easy detection and highly reliable and fast (30 fps) gesture classification.en
dc.format8 s.cs
dc.format.mimetypeapplication/pdf
dc.language.isoenen
dc.publisherVáclav Skala - UNION Agencyen
dc.relation.ispartofseriesWSCG '2015: short communications proceedingsen
dc.rights© Václav Skala - UNION Agencycs
dc.subjectrozpoznávání gest rukoucs
dc.subjectkinetikacs
dc.subjecthloubková mapacs
dc.subjecthistogramcs
dc.subjecthloubkové krájenícs
dc.subjectkonstelace otiskůcs
dc.titleRobust hand gesture recognition from 3D dataen
dc.typekonferenční příspěvekcs
dc.typeconferenceObjecten
dc.rights.accessopenAccessen
dc.type.versionpublishedVersionen
dc.subject.translatedhand gesture recognitionen
dc.subject.translatedkinecticsen
dc.subject.translateddepth mapen
dc.subject.translatedhistogramen
dc.subject.translateddepth slicingen
dc.subject.translatedfingertip constellationen
dc.type.statusPeer-revieweden
Vyskytuje se v kolekcích:WSCG '2015: Short Papers Proceedings

Soubory připojené k záznamu:
Soubor Popis VelikostFormát 
Kurmi.pdfPlný text4,1 MBAdobe PDFZobrazit/otevřít


Použijte tento identifikátor k citaci nebo jako odkaz na tento záznam: http://hdl.handle.net/11025/29678

Všechny záznamy v DSpace jsou chráněny autorskými právy, všechna práva vyhrazena.