Multimodal hand gesture recognition combining temporal and pose information based on CNN descriptors and histogram of cumulative magnitudes

Edwin Jonathan Escobedo Cardenas, Guillermo Camara Chavez

Research output: Contribution to journalArticle (Contribution to Journal)peer-review

28 Scopus citations

Abstract

In this paper, we present a new approach for dynamic hand gesture recognition. Our goal is to integrate spatiotemporal features extracted from multimodal data captured by the Kinect sensor. In case the skeleton data is not provided, we apply a novel skeleton estimation method to compute temporal features. Furthermore, we introduce an effective method to extract a fixed number of keyframes to reduce the processing time. To extract pose features from RGB-D data, we take advantage of two different approaches: (1) Convolutional Neural Networks and (2) Histogram of Cumulative Magnitudes. We test different integration methods to fuse the extracted spatiotemporal features to boost recognition performance in a linear SVM classifier. Extensive experiments prove the effectiveness and feasibility of the proposed framework for hand gesture recognition.

Original languageEnglish
Article number102772
JournalJournal of Visual Communication and Image Representation
Volume71
DOIs
StatePublished - Aug 2020
Externally publishedYes

Keywords

  • Convolucional neuronal networks
  • Fusion schemes
  • Hand gesture recognition
  • Histogram of cumulative magnitudes
  • Keyframe extraction
  • Pose and motion information
  • Spherical coordinates

Fingerprint

Dive into the research topics of 'Multimodal hand gesture recognition combining temporal and pose information based on CNN descriptors and histogram of cumulative magnitudes'. Together they form a unique fingerprint.

Cite this