Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Recognition method and system based on dual-modal emotion fusion of voice and facial expression

A facial expression and recognition method technology, applied in the field of emotion recognition, can solve the problems of poor facial expression recognition effect, great influence on emotion recognition results, and no consideration of acoustic channel obstruction, etc., to achieve excellent spatial position and direction selectivity, Effects of reduced computation, good rotation invariance, and grayscale invariance

Active Publication Date: 2019-12-17
CHINA UNIV OF GEOSCIENCES (WUHAN)
View PDF0 Cites 2 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0003] Existing emotion recognition is mainly based on single-modal emotion recognition. Most of the research on facial expression recognition is still on the emotion recognition of basic facial expressions, and the recognition effect on more subtle expressions is not good.
The research on speech emotion recognition is relatively mature, but when performing single-modal emotion recognition for speech, if the speech channel is blocked, the result of emotion recognition will be greatly affected
[0004] Ye Liang et al. proposed a speech feature screening method for mixed speech emotion recognition, which can well extract the best feature set in a series of acoustic features, but does not consider the situation where the acoustic channel is blocked
Zhao Xiaoming et al. proposed a robust speech emotion recognition method based on compressed sensing, which expanded the extraction of feature parameters from prosodic features and sound quality features to Mel frequency cepstral coefficient MFCC, which improved the anti-interference performance of feature signals, but still Unresolved emotion recognition when the acoustic channel cannot acquire the signal

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Recognition method and system based on dual-modal emotion fusion of voice and facial expression
  • Recognition method and system based on dual-modal emotion fusion of voice and facial expression
  • Recognition method and system based on dual-modal emotion fusion of voice and facial expression

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0065] The principles and features of the present invention are described below in conjunction with the accompanying drawings, and the examples given are only used to explain the present invention, and are not intended to limit the scope of the present invention.

[0066] In this embodiment, the data of the eNTERFACE'05 audio-video multimodal emotion database is used as the material, and the simulation platform is MATLAB R2015b.

[0067] Such as figure 1 As shown, the recognition method based on the bimodal emotional fusion of voice and facial expression includes the following steps:

[0068] S1. Obtain audio data and video data of an object to be identified;

[0069] S2. The audio data is preprocessed to obtain an emotional voice signal; the facial expression image is extracted from the video data, and the eyes, nose, and mouth regions are segmented, and images of three regions of a unified standard are obtained after preprocessing. ;

[0070] S3, extracting the voice emot...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention relates to a voice-and-facial-expression-based identification method for dual-modal emotion fusion. The method comprises: S1, audio data and video data of a to-be-identified object are obtained; S2, a face expression image is extracted from the video data and segmentation of an eye region, a nose region, and a mouth region is carried out; S3, a facial expression feature in each regional image is extracted from images of the three regions; S4, PCA analysis and dimensionality reduction is carried out on voice emotion features and the facial expression features; and S5, naive Bayesian emotion voice classification is carried out on samples of two kinds of modes and decision fusion is carried out on a conditional probability to obtain a final emotion identification result. According to the invention, fusion of the voice emotion features and the facial expression features is carried out by using a decision fusion method, so that accurate data can be provided for corresponding conditional probability calculation carried out at the next step; and an emotion state of a detected object can be obtained precisely by using the method, so that accuracy and reliability of emotion identification can be improved.

Description

technical field [0001] The invention belongs to the field of emotion recognition, and more specifically relates to a method and system for dual-mode emotion fusion and recognition based on speech and facial expressions. Background technique [0002] With the improvement of living standards and the rapid development of information technology, people's demand for intelligent life is getting higher and higher, and the ability of human-computer interaction is getting more and more attention. As an important part of human-computer interaction, affective computing has become a research topic. hotspot. At present, the research on emotion recognition at home and abroad is mainly divided into two categories, one is emotion recognition based on single modality, and the other is emotion recognition based on multi-modality. The main difference between the two types of emotion recognition is that multi-modal emotion recognition analyzes the information collected by multiple channels, an...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Patents(China)
IPC IPC(8): G10L15/02G10L15/06G10L15/08G10L25/63
CPCG10L15/02G10L15/063G10L15/08G10L25/63G10L2015/0631
Inventor 刘振焘吴敏曹卫华陈鑫潘芳芳徐建平张日丁学文
Owner CHINA UNIV OF GEOSCIENCES (WUHAN)
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products