Children speech emotion recognition method

A speech emotion recognition and emotion recognition technology, applied in speech recognition, speech analysis, speech synthesis, etc., can solve the problems of large emotion recognition, slow recognition speed, unfavorable speech recognition, etc., to improve the naturalness and improve the effect of human-computer interaction Effect

Inactive Publication Date: 2012-11-21
美育文化传媒(上海)有限公司
View PDF1 Cites 0 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

On the one hand, the method based on the assumption of framing increases the complexity of the algorithm and slows down the recognition speed. On the other hand, due to the differences between each individual's speech rate and pronunciation characteristics, framing based on time windows often destroys the relationship between speech and semantics. The overall coordination relationship is not conducive to speech recognition, especially not conducive to emotion recognition with a high degree of semantic correlation
[0006] At the same time, in the prior art, there is no research on the unique voice characteristics and emotional characteristics of children, and there is no method for emotional analysis of children's speech and emotional speech synthesis.

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Children speech emotion recognition method
  • Children speech emotion recognition method
  • Children speech emotion recognition method

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0021] Speech emotion recognition method of the present invention comprises setting up voice database, carries out emotion classifier training and three parts of speech emotion recognition, as Figure 1 to Figure 4 shown.

[0022] Among them, the establishment of the voice library is mainly for children's voice, including voice collection and classification.

[0023] Generally speaking, babies start to learn to speak when they are about 1 year old, and then their voices will change as they grow older, and there will be a stage of transition from a child's voice to an adult voice, that is, the "voice change period". "Period" occurs at different times, mostly between the ages of 12 and 17, and lasts from six months to a year. The vocal cords will change a lot in the "voice change period", so the child described in the present invention refers to a boy or girl before entering the voice change period. At the same time, in order to ensure the reliability of sampling, children who...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The present invention discloses a children speech emotion recognition method, comprising: creating a training voice database, classifier training and recognizing emotion. The classifier training comprises the step of extracting the speech emotion features, which includes performing pretreatment on the analog speech signal, including sampling, denoising and other basic signal treatments, performing HHT transformation on the pretreated emotion sentences, obtaining the Hilbert spectrum of the emotion speech; performing amplitude feature extraction, intonation feature extraction, tone feature extraction and rhythm feature extraction, and obtaining the emotion feature; reducing dimension of the emotion feature, and obtaining different quaternionic feature of emotion space; performing calculation on the sample set of the training speech database, obtaining the projection of a series of quaternionic features in national emotion state in the emotion space i, thus forming an emotion feature projection set; and carrying out classification training based on the SVM / HMM mixed system by using the data in the emotion feature projection set.

Description

technical field [0001] The invention relates to speech recognition, in particular to a speech emotion recognition method based on children's speech. Background technique [0002] At present, the smart electronic products for children on the market can play songs and sentences, and can also carry out simple conversations through the answering mode, but the sounds made by such electronic products and learning machines are very mechanical, monotonous and unnatural, making them only It can be used as a cold tool, and its humanization function is poor, so it cannot be fully accepted by children. The response mode of existing electronic products cannot make corresponding emotional and style adjustments according to children's emotional changes and children with different personalities. Can play pre-programmed answers with mechanical voices, so the human-computer interaction effect is not good. [0003] In the current research on speech emotion recognition, although there are meth...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
Patent Type & Authority Patents(China)
IPC IPC(8): G10L15/00G10L15/02G10L15/06G10L13/02G10L25/63
Inventor 李立志韩笑蕾贾晓光郭亮杰徐锡涛
Owner 美育文化传媒(上海)有限公司
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products