Emotion recognition method based on multimode voice information complementation and gate control
A speech information and emotion recognition technology, applied in character and pattern recognition, neural learning methods, biological neural network models, etc., can solve problems such as the proportion and balance of modal fusion representations that are rarely considered
- Summary
- Abstract
- Description
- Claims
- Application Information
AI Technical Summary
Problems solved by technology
Method used
Image
Examples
Embodiment Construction
[0017] The present invention will be further described below in conjunction with the accompanying drawings and specific embodiments, but not as a limitation of the present invention.
[0018] In the prior art, most speech emotion recognition models only consider the information of the speech modality but do not take the text, that is, its semantic information into account, and lack the balanced fusion of semantic information and audio information; and most current networks are often due to large Due to the impact of the scale pre-training model, the amount of parameters is huge, and it is difficult to implement it in some scenarios that require high real-time and lightweight.
[0019] The emotion recognition method based on multi-mode speech information complementary AND gate control provided by the present invention, such as figure 1 As shown, first, the audio features and text features in the target video are extracted. For the text mode, the pre-trained GloVe word embedding...
PUM
Login to View More Abstract
Description
Claims
Application Information
Login to View More 


