Video expression recognition method based on deep residual attention network

An expression recognition and attention technology, applied in neural learning methods, character and pattern recognition, biological neural network models, etc., can solve the problem of not taking into account the difference in the intensity of emotional representation of face images.

Pending Publication Date: 2020-10-20
TAIZHOU UNIV
View PDF1 Cites 11 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0007] The present invention aims to overcome the technology that the video expression recognition in the prior art does not take into account the difference in the intensity of emotion representation in each local area in the face image, and does not take into account the semantic gap between manual features and subjective emotions in the video The problem is to provide a video expression recognition method based on the deep residual attention network, which is realized by using the spatial attention mechanism. The w

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Video expression recognition method based on deep residual attention network
  • Video expression recognition method based on deep residual attention network
  • Video expression recognition method based on deep residual attention network

Examples

Experimental program
Comparison scheme
Effect test

Example Embodiment

[0048] Embodiment 1: A video expression recognition method based on a deep residual attention network of this embodiment, such as figure 1 As shown, including the following steps:

[0049] S1. Perform video data preprocessing on video samples;

[0050] Step S1 includes the following steps:

[0051] S1.1. Firstly, for each video sample, filter out the image frames of the peak intensity (apex) period;

[0052] S1.2. Use the haar-cascades detection model for face detection; the face detection in step S1.2 includes the following steps:

[0053] Step 1. First, convert the input picture into a grayscale image to remove color interference;

[0054] Step 2. Set the size of the search face box, search for faces in the input image in turn, and capture and save after finding the face;

[0055] Step 3. According to the standard distance between the two eyes, cut out images containing key expression parts such as the mouth, nose, and forehead from the original facial expression image, as the input of...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention discloses a video expression recognition method based on a deep residual attention network. The method comprises the following steps: S1, performing video data preprocessing on a video sample; S2, performing facial expression feature extraction on the face image by adopting a deep residual attention network; and S3, carrying out certain processing on the features extracted in the step S2, then carrying out training and testing, and outputting a final classification result of the facial expressions. The method is realized by adopting a spatial attention mechanism, weights in spatial distribution are generated for an input feature map, and thenweighted summation are performed with the feature map, so network learning is supervised to allocate different attention (weights) to different areas closely related to expressions in the face image, feature learning of a target area closely related to expressions in the face image can be focused, the feature representation capabilityof the deep residual network is improved, and the performance of video expression recognition is further improved.

Description

technical field [0001] The invention relates to the technical fields of image processing and pattern recognition, in particular to a video expression recognition method based on a deep residual attention network. Background technique [0002] The communication between people is full of emotion, the expression of emotion is the most primitive instinct of human beings, and the basic element of emotion is the aggregate of various expressions. In the past, people recorded their lives through words or photos. Now most of them record important memories and emotional expressions, such as emotions, emotions, and emotions, in the form of video blogs and short videos. [0003] Feature extraction is an important part of video expression recognition. In the early video expression recognition, most researchers used manual features for the classification of video expressions. Among them, representative manual features mainly include: Local Binary Pattern (LBP), Local Phase Quantization...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
IPC IPC(8): G06K9/00G06K9/62G06N3/04G06N3/08
CPCG06N3/084G06V40/168G06V40/174G06N3/045G06F18/214
Inventor 赵小明张石清
Owner TAIZHOU UNIV
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products