Methods for extracting and modeling Chinese speech emotion in combination with glottis excitation and sound channel modulation information

An extraction method and technology for modulating information, applied in the information field, can solve the problems of the complex emotional state of the speaker, the inability to completely separate the glottal excitation and the vocal tract modulation information, etc., to achieve the effect of improving the robustness

Inactive Publication Date: 2012-04-11
BEIHANG UNIV
View PDF4 Cites 15 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

It can be seen that the relationship between the content of the speech, the identity of the speaker, and the emotional state of the speaker is very complicated
[0003] Before the present invention, researchers mainly used various signal processing ...

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Methods for extracting and modeling Chinese speech emotion in combination with glottis excitation and sound channel modulation information
  • Methods for extracting and modeling Chinese speech emotion in combination with glottis excitation and sound channel modulation information
  • Methods for extracting and modeling Chinese speech emotion in combination with glottis excitation and sound channel modulation information

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0043] The technical solution of the present invention will be further elaborated below in conjunction with the accompanying drawings.

[0044] Such as figure 1 As shown, it is a flow chart of the extraction and recognition method of Chinese speech emotion points combined with glottal excitation and tone modulation, which is mainly divided into two parts: the extraction method of Chinese speech emotion points and the recognition method of Chinese speech emotion points.

[0045] One, the extracting method of Chinese glottal excitation information, the method steps are as follows:

[0046] Step 1. Formulate the electroglottogram emotional voice database specification;

[0047] Every step in the entire production process of the voice library should comply with specific specifications, including speaker specifications, recording script design specifications, recording specifications, audio file naming specifications, and experiment recording specifications. The production specific...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention discloses methods for extracting and modeling Chinese speech emotion in combination with glottis excitation and sound channel modulation information. The method for extracting Chinese glottis excitation information comprises the following steps of: setting an electric glottogram emotion speech database standard; collecting glottogram emotion speech data; and objectively evaluating glottogram emotion speech, i.e., evaluating the glottogram emotion speech with at least ten evaluators. The method for modeling Chinese emotion in combination with glottis excitation and sound channel modulation information comprises the following steps of: firstly, extracting glottis excitation information according to a glottogram; secondly, extracting sound channel modulation information in a speech signal according to the relation between the glottogram and the speech signal; and lastly, selecting an appropriate emotion characteristic for each emotion by using the Fisher rates of all standby characteristics, and training a one-to-many support vector machine model. In the invention, the glottis excitation information is extracted through a glottogram signal, and the glottis excitation information is separated from the sound channel modulation information in the speech signal, so that the emotion information in the speech can be reflected more accurately.

Description

(1) Technical field: [0001] The invention relates to a Chinese speech emotion extraction and modeling method combined with glottal excitation and vocal tract modulation information, belonging to the field of information technology. (two) background technology: [0002] The speech signal contains information such as speech content, speaker identity, and speaker's emotional state. Speech emotion recognition is to extract the speaker's emotional state from the speech signal. According to the excitation modulation model of speech, the speech signal is composed of glottal excitation and vocal tract modulation. Glottal excitation determines the change of speech prosody and plays an important role in speech emotion recognition. For example, when the speaker is emotional, the average value of the fundamental frequency of the speech signal will increase, and the range of the fundamental frequency will increase; when the speaker expresses doubts, the upward part of the sentence's fun...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
IPC IPC(8): G10L15/14
Inventor 毛峡陈立江
Owner BEIHANG UNIV
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products