Dynamic gesture recognition method based on hybrid deep learning model

A dynamic gesture, deep learning technology, applied in character and pattern recognition, instruments, computer parts, etc., can solve problems such as destroying 2D images, difficult training, loss of interactive information, etc.

Active Publication Date: 2017-07-28
BEIJING UNIV OF TECH
View PDF4 Cites 58 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

For high-dimensional data such as 2D and 3D, it is obvious that the vectorization operation will destroy the internal structure of 2D images, 3D videos, etc., resulting in the loss of hidden interactive information in the structure. In the traditional vector variable-based restricted Boltzmann machine research Based on this, Qi Guanglei et al. proposed a matrix variable-based restricted Boltzmann machine (Matrix Variable Restricted Boltzmann Machine, MVRBM) for high-dimensional data such as images. This model can better model 2D data, but the RBM and MVRBM models are unsupervised
In the field of deep learning, another model that has received wide attention is the Convolutional Neural Network (CNN). CNN has been successfully applied to various image analysis and understanding fields such as positioning, detection, and recognition classification. There are also methods based on CNN and Its variant 3DCNN models the time axis of the video sequence for video classification, which is a difficult and complex task with large model parameters, difficult to train, and requires a lot of training data

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Dynamic gesture recognition method based on hybrid deep learning model
  • Dynamic gesture recognition method based on hybrid deep learning model
  • Dynamic gesture recognition method based on hybrid deep learning model

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0053] The embodiment of the present invention provides a dynamic gesture recognition method. The invention faces the problem of dynamic gesture recognition and uses the advantages of CNN and MVRBM to design a method for pre-training a NN model based on a CNN-MVRBM hybrid model. This method integrates CNN's ability to express images and MVRBM's dimensionality reduction representation and pre-training capabilities for 2D signals, so as to achieve effective spatiotemporal representation of 3D dynamic gesture video sequences on the one hand, and improve the recognition performance of traditional NN on the other hand. .

[0054] The CNN-MVRBM-NN hybrid deep learning model includes two stages of training and testing. In the training phase, it combines the effective image feature extraction capabilities of CNN, the 2D signal modeling capabilities of MVRBM, and the supervised classification characteristics of NN. In the recognition stage, based on the aforementioned trained CNN and NN ...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention discloses a dynamic gesture recognition method based on a hybrid deep learning model. The dynamic gesture recognition method includes a training phase and a test phase. The training phase includes first, training a CNN based on an image set constituting a gesture video and then extracting spatial features of each frame of the dynamic gesture video sequence frame by frame using the trained CNN; for each gesture video sequence to be recognized, organizing the frame-level features learned by the CNN into a matrix in chronological order; inputting the matrix to an MVRBM to learn gesture action spatiotemporal features that fuse spatiotemporal attributes; and introducing a discriminative NN; and taking the MVRBM as a pre-training process of NN model parameters and network weights and bias that are learned by the MVRBM as initial values of the weights and bias of the NN, and fine-tuning the weights and bias of the NN by a back propagation algorithm. The test phase includes extracting and splicing features of each frame of the dynamic gesture video sequence frame by frame based on CNN, and inputting the features into the trained NN for gesture recognition. The effective spatiotemporal representation of the 3D dynamic gesture video sequence is realized by adopting the technical scheme of the invention.

Description

Technical field [0001] The invention belongs to the field of computer vision and machine learning, and specifically relates to a dynamic gesture recognition method based on a hybrid deep learning model. Background technique [0002] Gesture recognition has important applications in visual communication, human-computer interaction, and augmented reality. However, due to the complexity and variability of gestures and the influence of factors such as illumination and individual changes, the research on gesture recognition is still a challenging problem. Vision-based gesture recognition usually includes two aspects: feature extraction and classifier design. Commonly used classifiers include neural network (NN), hidden Markov model (HMM), etc. Gesture feature representation methods usually include: hand shape, hand center of gravity position, moment feature, scale-invariant feature transform (SIFT) feature, Histogram of Oriented Gradient (HOG) feature, etc. However, manually select...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
Patent Type & Authority Applications(China)
IPC IPC(8): G06K9/00G06K9/62
CPCG06V40/28G06V20/46G06F18/214
Inventor 施云惠淮华瑞李敬华王立春孔德慧尹宝才
Owner BEIJING UNIV OF TECH
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products