Lip language recognition method based on generative adversarial network and time convolutional network

A technology of convolutional network and recognition method, applied in the field of lip recognition based on generative adversarial network and temporal convolutional network, which can solve the problems of poor performance, influence of lip features, and the accuracy of attention mechanism needs to be improved.

Active Publication Date: 2021-05-18
HEBEI UNIV OF TECH
View PDF9 Cites 2 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

However, the lip language recognition task deals with a longer continuous image sequence, the context is more closely connected, and the span of the time dimension is larger, and the accuracy of the attention mechanism in the lip language recognition task still needs to be improved.
Another major difficulty in the lip language recognition task is that lip features are often affected by angle, lighting, and speaker identity, and feature extraction faces great uncertainty
Most of the recognition models use a feature extractor based on Residual Network (ResNet), which works well in laboratory conditions, but does not perform well when applied directly in the actual environment.

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Lip language recognition method based on generative adversarial network and time convolutional network
  • Lip language recognition method based on generative adversarial network and time convolutional network
  • Lip language recognition method based on generative adversarial network and time convolutional network

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0031] Specific examples of the present invention are given below. The specific embodiments are only used to further describe the present invention in detail, and do not limit the protection scope of the claims of the present application.

[0032] The present invention provides a kind of lip language recognition method (abbreviation method) based on generation confrontation network and time convolutional network, it is characterized in that, this method comprises the following steps:

[0033] S1, making raw data; the raw data includes recognition network raw data and dense multi-angle lip change raw data;

[0034] Preferably, in S1, making the original data of the recognition network is: obtaining the source video and subtitle files from the network through a Python web crawler, using the YOLOv5 face detection algorithm to obtain the face area in the source video, and then segmenting the face, And corresponding to the subtitle file, the original data of the identification net...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention discloses a lip language recognition method based on a generative adversarial network and a time convolutional network. According to the method, firstly, a lip deflection angle is judged through a ResNet angle classifier, then, a GAN two-stage converter is used for lip correction, and finally, input into TCN is carried out for feature recognition and classification to generate a lip language recognition result. According to the method, the influence of uncertainty of illumination intensity, illumination angle, recognition angle, speaker identity and the like in an actual environment, which cannot be solved by a traditional convolution model, on lip feature extraction is overcome, and the accuracy of lip language recognition is remarkably improved. According to the method, dense multi-angle lip change original data is designed, continuity of images of a single camera is achieved, continuity of lip images in an observation range is achieved to the maximum extent, the problem that an existing multi-angle model cannot process continuously-changing lip images in an actual environment is effectively solved, therefore, the lip language recognition precision is improved.

Description

technical field [0001] The invention belongs to the field of artificial intelligence and deep learning, in particular to a lip language recognition method based on a generative confrontation network and a temporal convolution network. Background technique [0002] With the development of science and technology and the improvement of hardware manufacturing level, the amount of information that computers can process is also increasing exponentially, which makes artificial intelligence technology based on deep learning enter a stage of rapid development, and artificial intelligence technology has become more and more It has been widely used in people's daily life, subtly changing people's production and lifestyle, and has become one of the indispensable and important technologies in human society. The application scenarios of artificial intelligence technology cover all aspects of production and life, including speech recognition, intelligent medical care, machine vision, intel...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
Patent Type & Authority Applications(China)
IPC IPC(8): G06K9/00G06K9/46G06K9/62G06N3/04G06N3/08
CPCG06N3/049G06N3/08G06V40/171G06V40/20G06V10/44G06N3/047G06N3/045G06F18/2415
Inventor 张成伟赵昊天张满囤齐畅崔时雨
Owner HEBEI UNIV OF TECH
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products