Gesture recognition method based on recurrent 3D convolutional neural network

A neural network and three-dimensional convolution technology, applied in the field of human-computer intelligent interaction, can solve the problems of difficult classification and achieve the effect of low equipment cost and convenient non-contact gesture recognition

Inactive Publication Date: 2018-01-16
BEIJING UNION UNIVERSITY
View PDF4 Cites 17 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

This application is based on electromyographic signals for gesture recognition. Due to the complexity of the cluster structure and t

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Gesture recognition method based on recurrent 3D convolutional neural network
  • Gesture recognition method based on recurrent 3D convolutional neural network
  • Gesture recognition method based on recurrent 3D convolutional neural network

Examples

Experimental program
Comparison scheme
Effect test

Example Embodiment

[0052] Example one

[0053] Such as figure 1 As shown, step 100 is performed to perform data preprocessing, which is used to process the acquired data into a fixed size to meet the input specification requirements of the input layer of the cyclic three-dimensional convolutional neural network. Such as figure 2 As shown, step 200 and step 210 are executed in sequence to crop the input data, and the randomly cropped size is A×A (in this embodiment, A=112). In step 220, in order to increase the diversity of the training samples, data enhancement is required. Random spatial rotation and scaling are performed on each video, the angle of spatial rotation is ±B° (in this embodiment, B=15), and the zoom size is ±C% (in this embodiment, C=20) . Step 230 is performed to perform random time scaling and jittering on each video, the scaling size is ±D% (in this embodiment, D=20), and the jitter amplitude is ±E frames (in this embodiment, E= 3). Step 240 is performed to obtain data that m...

Example Embodiment

[0059] Example two

[0060] Such as Figure 5 As shown, the overall system architecture consists of four parts: a data input module 500, a data preprocessing module 510, a cyclic three-dimensional convolutional neural network classifier 520, and an output class label 530. The cyclic three-dimensional convolutional neural network classifier 520 can be decomposed into: cyclic three-dimensional convolutional neural network classifier design sub-module 521, cyclic three-dimensional convolutional neural network classifier pre-training sub-module 522, cyclic three-dimensional convolutional neural network classifier training sub-module Module 523, cyclic three-dimensional convolutional neural network classifier optimization sub-module 524 and testing sub-module 525.

[0061] This embodiment proposes a method for gesture recognition based on a cyclic three-dimensional convolutional neural network, which includes importing video data in the data input module 500, performing data preprocessi...

Example Embodiment

[0062] Example three

[0063] Such as Image 6 As shown, step 600 is performed to collect image data through the camera (such as Figure 6a Shown). Step 620 is performed to crop the collected data, remove excess parts, and segment the gesture image (such as Figure 6b Shown). The image is divided into two stages for processing, namely the cyclic 3D convolutional neural network model training stage and the cyclic 3D convolutional neural network model testing stage. In the training phase of the cyclic three-dimensional convolutional neural network model, step 620 is executed to use the data enhancement technique to enhance the data, and the processing result is as follows Figure 6c Shown. Step 630 is executed to preprocess the data and extract clear key frames. Step 640 is performed to train the three-dimensional convolutional nerve first, and then train the overall model. Such as Figure 6d As shown, the method of training is to crop the video to obtain several pictures, and r...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention provides a gesture recognition method based on a recurrent 3D convolutional neural network, which includes a data preprocessing step. The method further includes the following steps: designing a recurrent 3D convolutional neural network classifier; pre-training the recurrent 3D convolutional neural network classifier; training the recurrent 3D convolutional neural network classifier;optimizing the recurrent 3D convolutional neural network classifier; outputting a classification label; and testing the classification effect of the classifier. Through the method, the effect of gesture recognition is more precise, the gestures in SKIG and ChaLearn2014 data sets can be accurately recognized, and the accuracy is higher than that of the traditional methods.

Description

technical field [0001] The invention relates to the technical field of human-computer intelligent interaction, in particular to a gesture recognition method based on a circular three-dimensional convolutional neural network. Background technique [0002] Gesture recognition is a topic in computer science and language technology. The purpose is to enable computers to recognize human gestures through mathematical algorithms, and to achieve effective communication and control between humans and computers. Gesture, with its advantages of intuitive, natural and easy-to-learn human-computer interaction and the realization of intelligent information exchange and processing without special hardware cooperation, has attracted more and more attention from enterprises and scholars. Among them, the current application Gesture recognition in the field of unmanned driving and VR is a simple gesture recognition. It is an artificial intelligence system that can understand the user's gestur...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
IPC IPC(8): G06K9/00G06K9/62
Inventor 袁家政刘宏哲邓智方
Owner BEIJING UNION UNIVERSITY
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products