Video emotion recognition method based on local enhanced motion history map and recursive convolutional neural network

A technology of motion history and local enhancement, applied in the field of pattern recognition, can solve the problems of low network classification ability, inability to make good use of video motion information, and small amount of data in facial expression video datasets, so as to prevent the small amount of training data, Improving the generalization ability and the effect of improving the classification ability

Active Publication Date: 2019-06-25
HEFEI UNIV OF TECH
View PDF10 Cites 9 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0007] However, studies have shown that for face-related problems such as facial expression recognition and face detection, using the original image as input does not make good use of the motion information in the video, making the network's classification ability not high
And because the amount of data in the expression video dataset is small, it is easy to make the network overfitting

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Video emotion recognition method based on local enhanced motion history map and recursive convolutional neural network
  • Video emotion recognition method based on local enhanced motion history map and recursive convolutional neural network
  • Video emotion recognition method based on local enhanced motion history map and recursive convolutional neural network

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0052] In this example, if figure 1 As shown, a video emotion recognition method based on local enhanced motion history graph and recursive convolutional neural network, including the following steps: obtain static expression picture data set and expression video data set, perform data expansion on video, and express expression video data set for preprocessing. A Local Enhanced Motion History Map (LEMHI) is then computed. Use the static image data set to pre-train the convolutional neural network (VGG16) model, the model structure is as follows figure 2 shown; then use LEMHI to fine-tune the pre-trained VGG16 model to obtain the LEMHI-CNN model. At the same time, the video frame is input into the pre-trained VGG16 model to extract spatial features, and the spatial features are stacked, sliced ​​and pooled to train the CNN-LSTM neural network model. Finally, the weighted fusion of the recognition results of the LEMHI-CNN model and the CNN-LSTM model is used to obtain the fi...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention discloses a video emotion recognition method based on a local enhanced motion history map and a recursive convolutional neural network, and the method comprises the steps: 1, obtaining astatic expression image data set and an expression video data set, and carrying out the preprocessing of the expression video data set; 2, calculating a local enhanced motion history map LEMHI; 3, pre-training a convolutional neural network VGG16 model by using the static picture data set; 4, performing fine tuning on the pre-trained VGG16 model by using LEMHI to obtain a LEMHI-CNN model; 5, inputting the video frame into a pre-trained VGG16 model to extract spatial features; 6, stacking, fragmenting and pooling the spatial features, and training an LSTM neural network model to obtain a CNN-LSTM model; 7, performing weighted fusion on the identification result of the LEMHI-CNN model and the CNN-LSTM model to obtain a final identification result. According to the invention, the video emotion recognition rate can be obviously improved.

Description

technical field [0001] The invention relates to a convolutional neural network, a cyclic neural network and classification discrimination, and belongs to the field of pattern recognition, in particular to a video emotion recognition method based on a dual-stream neural network. Background technique [0002] Traditional human-computer interaction, mainly through keyboards, mice, screens, etc., only pursues convenience and accuracy, and cannot understand and adapt to people's emotions and moods. Without this ability to understand and express emotions, it is difficult for a computer to have human-like intelligence. Emotion recognition is to endow computers with the ability to observe, understand and generate various emotional characteristics similar to humans, and finally enable computers to communicate and interact in a natural, friendly, and vivid way like humans. [0003] Research on video emotion recognition at home and abroad is generally divided into three steps: [000...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
Patent Type & Authority Applications(China)
IPC IPC(8): G06K9/00G06K9/62G06N3/04G06N3/08
Inventor 葛鹏胡敏王浩文王晓华任福继
Owner HEFEI UNIV OF TECH
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products