Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

A Dynamic Gesture Recognition Method Based on Hybrid Deep Learning Model

A dynamic gesture, deep learning technology, applied in the field of computer vision and machine learning, can solve problems such as loss of interactive information, large model parameters, and destruction of 2D images.

Active Publication Date: 2020-08-28
BEIJING UNIV OF TECH
View PDF4 Cites 0 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

For high-dimensional data such as 2D and 3D, it is obvious that the vectorization operation will destroy the internal structure of 2D images, 3D videos, etc., resulting in the loss of hidden interactive information in the structure. In the traditional vector variable-based restricted Boltzmann machine research Based on this, Qi Guanglei et al. proposed a matrix variable-based restricted Boltzmann machine (Matrix Variable Restricted Boltzmann Machine, MVRBM) for high-dimensional data such as images. This model can better model 2D data, but the RBM and MVRBM models are unsupervised
In the field of deep learning, another model that has received wide attention is the Convolutional Neural Network (CNN). CNN has been successfully applied to various image analysis and understanding fields such as positioning, detection, and recognition classification. There are also methods based on CNN and Its variant 3DCNN models the time axis of the video sequence for video classification, which is a difficult and complex task with large model parameters, difficult to train, and requires a lot of training data

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • A Dynamic Gesture Recognition Method Based on Hybrid Deep Learning Model
  • A Dynamic Gesture Recognition Method Based on Hybrid Deep Learning Model
  • A Dynamic Gesture Recognition Method Based on Hybrid Deep Learning Model

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0052] The embodiment of the present invention provides a dynamic gesture recognition method. The invention is oriented to the problem of dynamic gesture recognition, and utilizes the advantages of CNN and MVRBM to design a method for pre-training NN model based on CNN-MVRBM hybrid model. This method integrates CNN's ability to express images and MVRBM's dimensionality reduction representation and pre-training capabilities for 2D signals, so that on the one hand, it realizes an effective spatiotemporal representation of 3D dynamic gesture video sequences, and on the other hand, it improves the recognition performance of traditional NNs. .

[0053] The CNN-MVRBM-NN hybrid deep learning model includes two stages of training and testing. In the training phase, CNN's effective image feature extraction ability, MVRBM's ability to model 2D signals, and NN's supervised classification characteristics are combined. In the recognition stage, dynamic gesture recognition can be effective...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses a dynamic gesture recognition method based on a hybrid deep learning model, which includes: in the training stage, first training a CNN based on an image set constituting a gesture video, and then using the trained CNN to extract the spatial features of each frame of a dynamic gesture video sequence frame by frame; For each gesture video sequence to be recognized, the frame-level features learned by CNN are organized into a matrix in time order; the matrix is ​​input to MVRBM to learn the spatiotemporal features of gestures that fuse spatiotemporal attributes; a discriminative NN is introduced; Consider MVRBM as the pre-training process of NN model parameters, use the network weights and offsets learned by MVRBM as the initial values ​​of NN weights and offsets, and fine-tune the NN weights and offsets through the backpropagation algorithm; test In the stage, the features of each frame of the dynamic gesture video sequence are also extracted frame by frame based on CNN and stitched together, and input to the previously trained NN for gesture recognition. By adopting the technical scheme of the invention, effective spatiotemporal representation of 3D dynamic gesture video sequences is realized.

Description

technical field [0001] The invention belongs to the field of computer vision and machine learning, and in particular relates to a dynamic gesture recognition method based on a hybrid deep learning model. Background technique [0002] Gesture recognition has important applications in visual communication, human-computer interaction, augmented reality and other fields. However, due to the complexity and variability of gestures and the influence of factors such as illumination and individual changes, gesture recognition research is still a challenging problem. Vision-based gesture recognition usually includes two aspects: feature extraction and classifier design. Commonly used classifiers include neural network (NN), hidden Markov model (HMM), etc. Gesture feature representation methods usually include: hand shape, hand center of gravity position, moment feature, scale-invariant feature transform (Scale-invariant feature transform, SIFT) feature, histogram of oriented gradien...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Patents(China)
IPC IPC(8): G06K9/00G06K9/62
CPCG06V40/28G06V20/46G06F18/214
Inventor 施云惠淮华瑞李敬华王立春孔德慧尹宝才
Owner BEIJING UNIV OF TECH
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products