Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Continuous sign language statement recognition method based on modal matching

A sentence recognition and sign language technology, applied in the field of sign language recognition, can solve problems such as unsatisfactory practicability, time-consuming and labor-consuming, large parameter volume, etc., and achieve the effect of improving practical application ability, reducing high dependence, and reducing strict requirements

Active Publication Date: 2021-11-05
CHINA UNIV OF MINING & TECH
View PDF9 Cites 1 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0004] Sign language recognition based on wearable devices generally uses data gloves, sensors or some motion capture devices to collect hand movement data, extracts effective information from these data, and then models and classifies hand gestures. The recognition method can indeed accurately locate the hand, and the accuracy is high. However, data gloves are expensive, and sign language speakers must wear bulky gloves during demonstrations, which often make sign language demonstrators feel constrained; sign language recognition based on traditional machine learning is usually divided into two categories: Three steps, data preprocessing, feature extraction and modeling recognition, usually use algorithms such as scale invariant feature transformation and gradient direction histogram to design features manually, and then use traditional machine learning classifiers for modeling and recognition, such as SVM, HMM Such as traditional classifiers, sign language recognition based on traditional machine learning requires manual design of feature extraction methods, cannot automatically obtain image features, and relies on manual experience, which is time-consuming and labor-intensive
[0005] The feature extraction network structure is relatively complex, and training on a large-scale sign language data set will have a large amount of parameters and is very time-consuming, so that the task of continuous sign language sentence recognition cannot be completed in a targeted manner
In addition, the output sentence of sign language sentence recognition has the problem that it cannot conform to the daily grammatical relationship when the sentence structure is relatively complex, and it is difficult to train the codec network on a data set that lacks labels, which is not very practical.

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Continuous sign language statement recognition method based on modal matching
  • Continuous sign language statement recognition method based on modal matching
  • Continuous sign language statement recognition method based on modal matching

Examples

Experimental program
Comparison scheme
Effect test

Embodiment 1

[0076] The continuous sign language sentence recognition method based on modal matching of the present invention, the steps are as follows:

[0077] Step S1, collect 1000 sign language videos in color video mode, use the TV-L1 algorithm to extract the optical flow information of sign language videos in color video mode, and form an optical flow image sequence with the same number of frames, and provide optical flow video mode sign language video in different modalities; using CNN to extract the key frames of each sign language video in the above two modalities respectively, corresponding to the key frame sign language of the two modalities with a pixel size of 224×224 and uniform sampling of key frames to 8 frames For the video, 800 corresponding videos are selected from the key-frame sign language videos of the two modalities to form the training set, and the remaining videos in the key-frame sign language videos of the two modalities form the test set. The number D of key fr...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses a continuous sign language statement recognition method based on modal matching, and the method comprises the steps: taking a color video of a sign language statement, a key frame of an optical flow image sequence and a target word segment sequence as input, and, through a continuous sign language statement recognition model based on modal matching, matching and aligning the color video of the sign language statement and the key frame fragment sequence of the optical flow image sequence with semantics to obtain a final semantic sequence. The invention discloses a continuous sign language statement recognition model based on modal matching, which uses a lightweight feature extraction network to reduce the parameter quantity, performs task specific training for a sign language data set, and recognizes continuous sign language statements under the condition that the number of labels of samples in the data set is small; and the problem of high dependence on human body posture information in a traditional sign language recognition method is reduced.

Description

technical field [0001] The invention belongs to sign language recognition technology, in particular to a continuous sign language sentence recognition method based on modality matching. Background technique [0002] Sign language recognition uses computer technology to analyze the semantics of sign language used by humans, so that computers can understand sign language and convert sign language, a body language, into easy-to-understand text, voice and other forms of expression. [0003] With the rapid popularization of artificial intelligence, human-computer interaction technology has made human beings feel unprecedented convenience, which is of great significance in promoting the harmonious development of human society. For the first time, a continuous sign language sentence recognition method based on modal matching is proposed to align video and semantics, using color video and its corresponding optical flow video key frame sequence as input, enhancing data representation...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(China)
IPC IPC(8): G06K9/00G06K9/62G06N3/04
CPCG06N3/047G06N3/045G06N3/044G06F18/253
Inventor 王军袁静波申政文潘在宇李玉莲鹿姝
Owner CHINA UNIV OF MINING & TECH
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products