A Sign Language Word Recognition Method Based on Multimodal Hierarchical Information Fusion
A multi-modal and word-based technology, applied in the field of sign language recognition, can solve problems that ordinary people cannot afford, a large amount of labor costs, and affect sign language communication, and achieve the effect of improving sign language recognition methods
- Summary
- Abstract
- Description
- Claims
- Application Information
AI Technical Summary
Problems solved by technology
Method used
Image
Examples
Embodiment 1
[0051] combine figure 1 As shown, the present invention is a novel sign language word recognition method based on multimodal hierarchical information fusion, and the steps are as follows:
[0052] Step S1: Use the Kinect V2 depth camera to simultaneously collect color sign language video data, depth sign language video data, and skeletal node sign language video data of 60 commonly used sign language words to construct a multimodal Chinese daily sign language word dataset. The color video image resolution is 1920*1080, the depth video image resolution is 512*424, and the bone node video image resolution is 1920*1080.
[0053] Step S2: Use the CNN network to extract and collect key frames in the video sequence, and after obtaining the key frames, cut the T frame key frame data into a network input size map N*N size (N=224), and then normalize the image data deal with.
[0054] Step S3: Input the preprocessed T frames of color video key frame data and T frames of depth video k...
PUM
Login to View More Abstract
Description
Claims
Application Information
Login to View More 


