Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Real-time video face key point detection method based on deep learning

A face key point, real-time video technology, applied in neural learning methods, instruments, biological neural network models, etc., can solve problems such as inability to use global inter-frame information, poor real-time performance, and poor detection accuracy for large face poses.

Active Publication Date: 2021-05-14
HEBEI UNIV OF TECH
View PDF6 Cites 10 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0005] The present invention aims at the shortcomings of the face key point detection algorithm currently applied to video, such as poor real-time performance, inability to use global inter-frame information, and poor detection accuracy for large gestures of the face, and proposes a real-time video face key point based on deep learning Detection method

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Real-time video face key point detection method based on deep learning
  • Real-time video face key point detection method based on deep learning
  • Real-time video face key point detection method based on deep learning

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0027] The present invention will be further described below in conjunction with the accompanying drawings, but the protection scope of the present invention is not limited thereto.

[0028] The method for detecting key points of real-time video faces based on deep learning in the present invention mainly includes steps such as constructing a convolutional neural network for key point detection, single-frame model training, cross-frame smooth training, and frame-by-frame detection.

[0029] In the construction of the key point detection convolution network, based on the SAN network, the ordinary convolution is replaced by the depth separable convolution, the network structure is lightweight, and the boundary heat map subtask is added to improve the accuracy of the model detection. Figure, the loss function to solve the imbalanced distribution of face pose samples;

[0030]Perform single-frame model training, including the following steps:

[0031] Read the training samples in...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention relates to a real-time video face key point detection method based on deep learning, and the method employs a convolutional neural network to carry out the key point detection of a single frame, employs a depth separable convolution to improve the model detection rate, employs a boundary heat map as an additional subtask of an original network to improve the constraint of a global face structure of the original network. The method improves the detection accuracy of an original network, is used for solving a data imbalance loss function of a heat map, improves the generalization capability of a model for a large attitude sample under a limited sample, and improves the inter-frame smoothness through an optical flow loss function. In the detection process, for a frame of which the confidence is lower than a key point confidence threshold due to an extremely large angle, fitting is carried out by utilizing 3DMM to obtain dense key point coordinates, 68-point sampling is carried out on the obtained dense key points according to a projection error between minimum frames, and the consistency with the previous frame is kept. The method has the advantages of real-time performance, capability of utilizing global inter-frame information, high detection accuracy of a face large posture condition and the like.

Description

technical field [0001] The technical solution of the present invention relates to face key point detection, in particular to a real-time video face key point detection method based on deep learning. Background technique [0002] Face key point detection is to detect the coordinates of key points such as eyebrows, eyes, nose, etc. for the input face area. Face key point detection plays a key role in the practical application of face-related research in face recognition, 3D face reconstruction, facial expression cloning, and micro-expression analysis. [0003] Face key point detection in images is easily affected by background, face pose, dramatic expression changes, occlusion, lighting, etc., while video key point detection also has high requirements for inter-frame smoothness and real-time performance in addition to the above. Existing key point detection methods for video often only use optical flow to track local key points to improve speed, and cannot use global inter-fr...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
IPC IPC(8): G06K9/00G06K9/46G06N3/04G06N3/08
CPCG06N3/08G06V40/171G06V10/462G06N3/045
Inventor 张满囤齐畅崔时雨刘川申冲权子洋师子奇
Owner HEBEI UNIV OF TECH
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products