Speech emotion recognition method based on semi-supervised feature selection

A technology for speech emotion recognition and feature selection, applied in speech analysis, instruments, etc., can solve the problems of difficulty in labeling samples, feature differences, and speaker differences.

Active Publication Date: 2014-08-27
SOUTH CHINA UNIV OF TECH
View PDF5 Cites 14 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

First of all, the popular structure of data plays a very important role in speech emotion recognition. This is because the differences between speakers are relatively large, and it is difficult for labeled samples of people whose emotions are to be recognized to appear in the training data. In the algorithm, only the category structure of the data is considered for feature selection, then the selected features will overfit the training data and have poor recognition ability for new test samples
Secondly, the information provided by the unlabeled samples of the people whose emotions are to be recognized also plays a very important role. Also because the di

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Speech emotion recognition method based on semi-supervised feature selection
  • Speech emotion recognition method based on semi-supervised feature selection
  • Speech emotion recognition method based on semi-supervised feature selection

Examples

Experimental program
Comparison scheme
Effect test

Embodiment 1

[0043] Such as figure 1 As shown, the speech emotion recognition method based on semi-supervised feature selection described in this embodiment includes two parts, the training phase and the recognition phase, now combined figure 1 The flow charts are detailed below.

[0044] 1. Training stage

[0045] In this stage, all speakers are trained separately to obtain the classifier corresponding to each speaker. The specific process is as follows:

[0046] Step 1: Extract MFCC, LFPC, LPCC, ZCPA, PLP, R for all speech training signals (for each training session, the speech signals of all labeled samples and the speech signal of a current speaker's unlabeled sample) -PLP features, where the number of Mel filters of MFCC and LFPC is 40; the linear prediction orders of LPCC, PLP, and R-PLP are 12, 16, and 16 respectively; the frequency segments of ZCPA are: 0, 106, 223, 352, 495, 655, 829, 1022, 1236, 1473, 1734, 2024, 2344, 2689, 3089, 3522, 4000. Therefore, the dimensions of each...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention discloses a speech emotion recognition method based on semi-supervised feature selection. According to the method, a specific classifier is trained for each speaker, so that the negative influence of speaker difference on speech emotion recognition is reduced. The training method comprises the steps of extracting the features of a label sample and a no-label sample of a certain speaker, obtaining the statistic result of all the features by means of multiple statistic functions, and executing the normalization algorithm; selecting a feature which can highlight the speech emotion of the speaker to be tested by means of the semi-supervised feature selection algorithm, wherein the semi-supervised feature selection algorithm can consider the manifold structure of data, the classification structure of data and information provided through the no-label data of the speaker to be tested at the same time; finally, training the classifier for recognition of speech emotion of the speaker to be tested by means of a support vector machine. By the adoption of the method, high recognition accuracy can be realized when the sample number for the speaker normalization algorithm is small.

Description

technical field [0001] The invention relates to the research fields of speech signal processing and pattern recognition, in particular to a speech emotion recognition method based on semi-supervised feature selection. Background technique [0002] With the continuous development of information technology, social development puts forward higher requirements for affective computing. For example, in terms of human-computer interaction, a computer with emotional capabilities can acquire, classify, recognize, and respond to human emotions, thereby helping users to obtain an efficient and friendly feeling, and can effectively reduce people's frustration in using computers, and even Can help people understand their own and other people's emotional world. For example, using such technology to detect whether the driver is focused, the level of stress he feels, etc., and respond accordingly. In addition, affective computing can also be applied in related industries such as robots, s...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
IPC IPC(8): G10L25/63G10L17/14G10L17/02G10L17/04
Inventor 文贵华孙亚新
Owner SOUTH CHINA UNIV OF TECH
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products