Unlock instant, AI-driven research and patent intelligence for your innovation.

Multi-modal emotion recognition method based on feature-time attention mechanism

A technology of emotion recognition and attention, applied in the field of pattern recognition, can solve the problems of low relative accuracy and less information, and achieve the effect of improving accuracy

Active Publication Date: 2021-05-11
SOUTHEAST UNIV
View PDF3 Cites 20 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

Decision fusion is to make a final judgment on the final result with certain rules after the emotion recognition results are obtained by each modal model, which has high flexibility and strong real-time performance. Less volume, lower relative accuracy

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Multi-modal emotion recognition method based on feature-time attention mechanism
  • Multi-modal emotion recognition method based on feature-time attention mechanism
  • Multi-modal emotion recognition method based on feature-time attention mechanism

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0082] Below in conjunction with accompanying drawing, technical scheme of the present invention is described in further detail:

[0083] This invention may be embodied in many different forms and should not be construed as limited to the embodiments set forth herein. Rather, these embodiments are provided so that this disclosure will be thorough and complete, and will fully convey the scope of the invention to those skilled in the art.

[0084] Such as figure 1 As shown, it is a schematic diagram of the overall flow of a multi-modal emotion recognition method based on a feature-temporal attention mechanism proposed by the present invention. The method specifically includes the following steps,

[0085] Step 1: Build an emotion recognition network model, obtain audio and video samples containing emotional information, extract face grayscale images from the video modality data in the samples, and encode them into fixed-dimensional feature vectors using a deep residual network ...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses a multi-modal emotion recognition method based on a feature-time attention mechanism, and the method comprises the following steps: constructing a neural network model, and obtaining an audio / video sample containing emotion information and a video primary feature matrix; acquiring an audio primary feature matrix; obtaining a fusion feature matrix according to the video primary feature matrix and the audio primary feature matrix, and inputting the fusion feature matrix into a feature self-attention mechanism module; inputting the processed fusion feature matrix into a bidirectional gating cycle unit network to obtain output vectors at all moments and a state vector of the last hidden layer; obtaining an attention weight, and obtaining an advanced feature vector according to the attention weight; obtaining a neural network model capable of performing sentiment classification on audio and video samples after training; and collecting to-be-detected audios and videos, and inputting the to-be-detected audios and videos into the trained neural network model to obtain an emotion classification result. According to the invention, the accuracy of face emotion recognition in audios and videos can be improved.

Description

technical field [0001] The invention relates to the technical field of pattern recognition, in particular to a multimodal emotion recognition method based on a feature-time attention mechanism. Background technique [0002] As an important basis of human life experience, emotion affects human cognition, perception and daily life. In 1971, psychologists Ekman and Friesen divided human emotions into six basic emotional categories through cross-cultural research, which are happy (Happy), sad (Sad), surprised (Surprise), angry (Angry), fear ( Fear) and Disgust (Disgust), these six types of emotion categories are universal, and more fine-grained sub-level emotion categories can be synthesized on this basis. In 1997, Professor Picard first proposed the concept of "affective computing", which involves psychology, cognition, pattern recognition, speech signal processing, physiology, sociology, computer vision, and artificial intelligence. Facial expressions, voice and other inform...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(China)
IPC IPC(8): G06K9/00G06K9/62G06N3/04G06N3/08
CPCG06N3/084G06V40/174G06V40/161G06N3/047G06N3/045G06F18/214G06F18/253
Inventor 李克梁瑞宇赵力郭如雪
Owner SOUTHEAST UNIV