Unlock instant, AI-driven research and patent intelligence for your innovation.

A Sign Language Word Recognition Method Based on Multimodal Hierarchical Information Fusion

A multi-modal and word-based technology, applied in the field of sign language recognition, can solve problems that ordinary people cannot afford, a large amount of labor costs, and affect sign language communication, and achieve the effect of improving sign language recognition methods

Active Publication Date: 2022-03-18
CHINA UNIV OF MINING & TECH
View PDF22 Cites 0 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

However, manual translation requires a lot of labor costs, which can only be used in formal occasions, and ordinary people cannot afford it; the way of wearing equipment will bring equipment burden to sign language users, and the limitation of equipment will affect normal sign language communication

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • A Sign Language Word Recognition Method Based on Multimodal Hierarchical Information Fusion
  • A Sign Language Word Recognition Method Based on Multimodal Hierarchical Information Fusion
  • A Sign Language Word Recognition Method Based on Multimodal Hierarchical Information Fusion

Examples

Experimental program
Comparison scheme
Effect test

Embodiment 1

[0051] combine figure 1 As shown, the present invention is a novel sign language word recognition method based on multimodal hierarchical information fusion, and the steps are as follows:

[0052] Step S1: Use the Kinect V2 depth camera to simultaneously collect color sign language video data, depth sign language video data, and skeletal node sign language video data of 60 commonly used sign language words to construct a multimodal Chinese daily sign language word dataset. The color video image resolution is 1920*1080, the depth video image resolution is 512*424, and the bone node video image resolution is 1920*1080.

[0053] Step S2: Use the CNN network to extract and collect key frames in the video sequence, and after obtaining the key frames, cut the T frame key frame data into a network input size map N*N size (N=224), and then normalize the image data deal with.

[0054] Step S3: Input the preprocessed T frames of color video key frame data and T frames of depth video k...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses a sign language word recognition method based on multi-modal hierarchical information fusion. The method includes: taking the key frame sequence of three modes of color video, depth video and skeleton node video as network input, and constructing a dual-stream I3D network extracts color video and depth video features, integrates dual-modal semantic features through feature splicing, and then uses LSTM to construct long-term spatio-temporal features, and uses SoftMax to classify and score; at the same time, use DST-GCN network to extract spatio-temporal features of skeletal node videos, and then use SoftMax performs classification and scoring; finally, the prediction scores of the two SoftMax layers are fused through decision-level fusion to obtain sign language word recognition results. The sign language word recognition method based on multi-modal hierarchical information fusion proposed by the present invention fully utilizes the complementary information of multi-modal data by constructing a hierarchical fusion strategy; by constructing a DST‑GCN network, the spatio-temporal map volume is enhanced Accumulated network time feature extraction ability, thereby improving the accuracy of sign language word recognition.

Description

technical field [0001] The invention belongs to the technical field of sign language recognition, and specifically refers to a sign language word recognition method based on multimodal hierarchical information fusion. Background technique [0002] According to data, the number of deaf-mute people in my country has exceeded 20.8 million, accounting for about 1.69% of the total population of our country. In addition, there are a large number of people suffering from hearing impairment. For them, sign language is their communication medium, but only a few people have mastered sign language and can communicate with it, which makes deaf people have many communication barriers in their lives. [0003] At present, the main solutions are: manual-based translation and sign language users wearing specific equipment on their hands for translation. However, manual translation requires a lot of labor costs, which can only be used in formal occasions, and ordinary people cannot afford it...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Patents(China)
IPC IPC(8): G06V40/20G06V10/80G06V10/764G06V10/82G06V10/62G06K9/62G06N3/04
CPCG06V40/28G06N3/044G06N3/045G06F18/2415G06F18/253
Inventor 王军吕智成申政文李玉莲潘在宇鹿姝
Owner CHINA UNIV OF MINING & TECH