Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Sign language video translation method based on fusion of temporal convolutional network and recurrent neural network

A technology of recurrent neural network and convolutional network, applied in natural language translation, instruments, computing, etc., can solve the problems of ignoring related information and changing information, affecting the effect of recognition, and difficult to learn videos.

Active Publication Date: 2019-10-18
HEFEI UNIV OF TECH
View PDF12 Cites 0 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

The disadvantage of these features is that when extracting features, we only pay attention to the information of each video frame, while ignoring the correlation information and change information between frames in continuous video, which will affect the effect of subsequent recognition.
[0005] In the learning process of sequence models, commonly used models include support vector machines, dynamic time warping algorithms, hidden Markov models, and other traditional models. These models are more suitable for recognizing and translating single sign language words, and for continuous and multiple Videos of gestures are difficult to learn, and it is also difficult to learn the semantic correspondence between actions and words, so it is impossible to effectively translate continuous sign language sentences

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Sign language video translation method based on fusion of temporal convolutional network and recurrent neural network
  • Sign language video translation method based on fusion of temporal convolutional network and recurrent neural network
  • Sign language video translation method based on fusion of temporal convolutional network and recurrent neural network

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0051] In this example, if figure 1 As shown, a sign language video translation method based on the fusion of time-domain convolutional network and cyclic neural network is to fully extract the spatial and temporal features in sign language video, effectively learn the features of key actions with high recognition, and effectively Avoid being interfered by factors such as the signer's body shape, sign language speed, and sign language habits during the model learning process. Its steps include:

[0052] First, the original sign language video is preprocessed to extract sign language video features; then two different network structures (time-domain convolutional network TCN and bidirectional recurrent neural network BGRU) are used simultaneously to encode continuous sign language video features and output each slice word Generated probability; then the output of the middle layer of the sub-network is spliced ​​and sent to the fusion network (FL) to learn and generate a word s...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses a continuous sign language video translation method based on the fusion of time-domain convolutional network and cyclic neural network, comprising the following steps: feature extraction of sign language video and construction of word list; processing of time-domain convolutional network TCN; two-way loop The processing of neural network BGRU; the word mapping process of features; the processing of fusion network FL; the optimization of network model parameters based on the fusion of time-domain convolutional network TCN and bidirectional recurrent neural network BGRU; fusion and decoding of word encoding vectors. The present invention can use network structures from different perspectives to effectively overcome communication barriers caused by inaccurate interpretation of sign language translation, use different networks to study and analyze different forms of data, and further improve the accuracy of sign language translation , to increase the robustness of sign language translation.

Description

technical field [0001] The invention belongs to the technical field of computer vision, and relates to technologies such as pattern recognition, natural language processing, and artificial intelligence. Specifically, it relates to a continuous sign language video translation method based on the fusion of time-domain convolutional networks and cyclic neural networks. Background technique [0002] Sign language is a way for normal people to communicate with deaf-mute people. It usually consists of a series of meaningful movements consisting of the sign language user's body movements, joint movements and facial expressions. However, there are often communication barriers between normal people who have not learned sign language and sign language users. Therefore, how to capture information such as gestures of sign language users and convert them into information that normal people can understand has gradually attracted people's attention. A good sign language translation system...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Patents(China)
IPC IPC(8): G06K9/00G06F17/28
CPCG06F40/58G06V40/28
Inventor 郭丹王硕汪萌
Owner HEFEI UNIV OF TECH
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products