Cross-database Speech Emotion Recognition Method and Device Based on Joint Distribution Least Squares Regression
A technology of speech emotion recognition and least squares, applied in speech analysis, instruments, etc., to achieve good adaptability and accurate recognition results
- Summary
- Abstract
- Description
- Claims
- Application Information
AI Technical Summary
Problems solved by technology
Method used
Image
Examples
Embodiment Construction
[0064] This embodiment provides a cross-database speech emotion recognition method based on joint distribution least squares regression, such as figure 1 shown, including the following steps:
[0065] (1) Acquire two speech databases as training database and test database, wherein the training speech database contains several speech fragments and corresponding speech emotion category labels, while the test database only contains several speech fragments to be recognized.
[0066] In this embodiment, we use three types of speech emotion databases commonly used in emotional speech recognition: Berlin, eNTERFACE and CAISA. Because the three types of databases contain different sentiment categories, the data are selected in the pairwise comparison. When comparing Berlin and eNTERFACE, we selected 375 pieces of data and 1077 pieces of data respectively, and the emotion categories were 5 categories (angry, scared, happy, disgusted, sad); when Berlin and CAISA were compared, we sele...
PUM
Login to View More Abstract
Description
Claims
Application Information
Login to View More 


