Unlock instant, AI-driven research and patent intelligence for your innovation.

Video expression recognition method based on C3D-SA

A C3D-SA, facial expression recognition technology, applied in the field of neural network and computer vision, can solve problems such as inability to take into account video motion information, poor video sequence processing effect, and insufficient intelligence in traditional feature extraction.

Active Publication Date: 2022-08-05
SHAANXI NORMAL UNIV
View PDF4 Cites 1 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0004] 2D CNN can perform different levels of spatial feature extraction on a single static image, but it cannot take into account the motion information between frames in the video, so it is not effective for video sequence processing
[0008] The above methods can recognize video facial expressions to a certain extent, but there are deficiencies: 1) The traditional feature extraction is not intelligent enough, the algorithm processing data is not efficient, and the performance is weak; 2) 3D CNN can only extract local spatio-temporal information and cannot Learn about long-term dependencies

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Video expression recognition method based on C3D-SA
  • Video expression recognition method based on C3D-SA
  • Video expression recognition method based on C3D-SA

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0023] see figure 1 , in one embodiment, it discloses a C3D-SA-based video expression recognition method, comprising the following steps:

[0024] S100: extracting expression features from a video sequence through a three-dimensional convolutional neural network to obtain an expression feature matrix;

[0025] S200: Connect the self-attention mechanism layer to learn the correlation between the features in the expression feature matrix, obtain an attention weight value, and then weight the expression feature matrix to obtain a weighted expression feature matrix;

[0026] S300: Connect the global mean pooling layer to perform feature mapping and dimension reduction on the weighted expression feature matrix, and then randomly discard some of the eigenvalues ​​in the weighted expression feature matrix through the loss layer to obtain a new expression feature matrix ;

[0027] S400: Connect the fully connected layer to perform feature mapping on the new expression feature matrix...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

A C3D-SA-based video expression recognition method comprises the steps of S100, performing expression feature extraction on a video sequence through a three-dimensional convolutional neural network to obtain an expression feature matrix; s200, connecting the correlation between features in the self-attention mechanism layer learning expression feature matrix to obtain an attention weight value, and weighting the expression feature matrix to obtain a weighted expression feature matrix; s300, connecting a global mean pooling layer to perform feature mapping and dimension reduction on the weighted expression feature matrix, and randomly discarding part of feature values in the weighted expression feature matrix through a loss layer to obtain a new expression feature matrix; and S400, connecting a full connection layer to carry out feature mapping on the new expression feature matrix to obtain a final feature matrix, and outputting an expression recognition label from the obtained final feature matrix through a softmax layer. According to the method, the 3D convolutional neural network and the self-attention mechanism are combined to improve the recognition accuracy of the video facial expression.

Description

technical field [0001] The present disclosure belongs to the technical fields of computer vision and neural networks, and particularly relates to a video expression recognition method based on C3D-SA. Background technique [0002] Facial expression recognition is one of the research hotspots in the fields of computer vision, pattern recognition and human emotion understanding, and plays an increasingly important role in public security, criminal investigation, medical care, education, retail and other fields. [0003] Convolutional Neural Networks (CNN) is a classic deep learning neural network, which is widely used in facial expression recognition. Multilayer Perceptron (MLP) is more robust. The basic layer of the convolutional neural network includes the convolutional layer, the pooling layer, the activation layer and the fully connected layer, of which the convolutional layer is the most important part, which contains multiple convolution kernels for convolution operatio...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(China)
IPC IPC(8): G06V40/16G06V10/82G06V10/764G06V10/42G06N3/08G06N3/04
CPCG06V40/174G06V10/435G06V10/764G06V10/82G06N3/08G06N3/047G06N3/045
Inventor 吴燕妮姚若侠范虹
Owner SHAANXI NORMAL UNIV