Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

A Mute Language Action Recognition Method Based on Two-way 3dCNN Model

A technology of action recognition and modeling, applied in the field of computer vision, to achieve the effect of improving the quality of life and spiritual well-being index

Active Publication Date: 2022-04-12
NORTHEASTERN UNIV LIAONING
View PDF5 Cites 0 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0003] The technical problem to be solved by the present invention is to provide a method for dumb language action recognition based on a two-way 3DCNN model, which can correctly capture gestures and body information of the human body, and also adds time dimension information. Coherent sentences can be obtained, and the customer service can only recognize the problem of a single vocabulary

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • A Mute Language Action Recognition Method Based on Two-way 3dCNN Model
  • A Mute Language Action Recognition Method Based on Two-way 3dCNN Model
  • A Mute Language Action Recognition Method Based on Two-way 3dCNN Model

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0029] The specific implementation manners of the present invention will be further described in detail below in conjunction with the accompanying drawings and embodiments. The following examples are used to illustrate the present invention, but are not intended to limit the scope of the present invention.

[0030] The traditional convolutional neural network can only obtain the spatial features of the input data, but for sign language videos, the features in the time dimension are also very important. Therefore, this method hopes to use the 3DCNN model framework to simultaneously extract the dumb language video streams. 3D here does not refer to the 3D in three-dimensional space, but refers to the 3D data formed after introducing the time dimension on the two-dimensional image, that is, the data composed of a series of video frames. At the same time, dumb language movements are different from general gestures. In addition to the most important hand information, dumb language ...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention provides a dumb language action recognition method based on a two-way 3DCNN model, and relates to the technical field of computer vision. The steps of the present invention are as follows: step 1: collect video data sets; step 2: preprocess the video data sets; step 3: set up a 3D convolutional neural network model for extracting gesture local information, and output a set of feature vectors of hands; step 4: Establish a 3D convolutional neural network model that extracts overall global information, and extract a set of overall global feature vectors; Step 5: Establish a two-way 3D convolutional neural network model, and obtain a feature map with local gesture information and global overall information ; Step 6: Obtain the word embedding feature vector, input the feature map and feature vector into the long-short-term memory network that generates dumb sentences, and obtain coherent sentences corresponding to dumb actions through iterative training. This method can correctly capture the gestures and body information of the human body to obtain coherent sentences, and overcomes the problem of only recognizing a single vocabulary.

Description

technical field [0001] The present invention relates to the technical field of computer vision, in particular to a method for recognizing dumb language actions based on a two-way 3DCNN model. Background technique [0002] Mute language is the most natural way for deaf people to communicate and communicate, and it is also an important way for deaf schools to teach and convey ideas. There are currently about 5,500 routine gestures included in Chinese textbooks for the dumb language, each gesture corresponding to a Chinese word. The purpose of dumb language recognition is to provide an effective and accurate mechanism through the computer to translate the dumb language into text or speech to make the communication between deaf-mute and hearing people more convenient and fast. When the deaf-mute communicates with the outside world, if the other party does not understand dumb language, it will cause a lot of trouble. Therefore, many experts and scholars have carried out a series...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Patents(China)
IPC IPC(8): G06V40/20G06V20/40G06V10/32G06V10/44G06V10/764G06V10/82G06K9/62G06N3/04G06N3/08
CPCG06N3/08G06V40/28G06V20/40G06V10/32G06V10/44G06N3/045G06F18/2414
Inventor 王斌杨晓春赵征
Owner NORTHEASTERN UNIV LIAONING
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products