Unlock instant, AI-driven research and patent intelligence for your innovation.

Bimodal fusion emotion recognition method based on biological radar and voice information

A voice information and emotion recognition technology, applied in the field of emotion recognition, can solve the problems of low accuracy of emotional features, complex and changeable human emotions, etc., and achieve the effect of improving the degree of freedom of measurement

Pending Publication Date: 2021-07-23
CHANGCHUN UNIV OF SCI & TECH
View PDF3 Cites 1 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0004] However, the emotions of the human body are complex and changeable, and the accuracy of judging the emotional characteristics by measuring one of the information alone is low. In order to improve the accuracy, the present invention is hereby proposed

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Bimodal fusion emotion recognition method based on biological radar and voice information
  • Bimodal fusion emotion recognition method based on biological radar and voice information
  • Bimodal fusion emotion recognition method based on biological radar and voice information

Examples

Experimental program
Comparison scheme
Effect test

Embodiment 1

[0053] A dual-modal fusion emotion recognition method based on bio-radar and voice information, such as figure 1 described, including the following steps:

[0054] 1) Acquisition of speech signals and physiological signals: non-contact collection of natural speech and human body signals using microphones and radars;

[0055] Radar refers to frequency modulation continuous wave radar, which uses linear frequency modulation technology to transmit sawtooth waves;

[0056] The microphone refers to a digital MEMS microphone that outputs a 1 / 2 cycle pulse density modulated digital signal;

[0057]2) Signal preprocessing: Preprocess the signals of the two modalities of physiology and speech, including heartbeat signals, breathing signals, and speech signals, so that they meet the input requirements of the corresponding models of different modalities;

[0058] 3) Emotional feature extraction: perform feature extraction on the preprocessed heartbeat signal, breathing signal, and voic...

Embodiment 2

[0062] Physiological information processing flow, such as figure 2 described, including the following steps:

[0063] 1) Use frequency-modulated continuous wave radar to obtain human physiological sign signals, perform band-pass filtering on the original human physiological sign signals, filter out drift signals below 0.2 Hz and noise signals above 2 Hz in the original sign signals, and convert 0.2 Hz-0.9 Hz signals to It is classified as a respiratory signal, and the signal of 0.9Hz-2.0Hz is classified as a heartbeat signal;

[0064] 2) Extract the time feature, waveform feature and frequency domain feature of the respiratory signal; extract the time feature, waveform feature and frequency domain feature of the heartbeat signal;

[0065] 3) PCA dimensionality reduction is carried out on the extracted respiratory signal and heartbeat signal features, and reduced to two-dimensional data to obtain physiological features;

[0066] 4) Input the physiological feature data to be ...

Embodiment 3

[0068] Voice information processing flow, such as image 3 described, including the following steps:

[0069] 1) Utilize the digital MEMS microphone to obtain the human voice signal, and pre-emphasize the human voice signal through a digital filter, and output the pre-emphasized voice data;

[0070] 2) utilizing short-term analysis technology to carry out frame processing to the voice data after the pre-emphasis, and obtain the time series of voice feature parameters;

[0071] 3) Using the Hamming window function to perform windowing processing on the speech feature parameter time series to obtain speech windowing data

[0072] 4) utilize double-threshold comparison method to carry out endpoint detection to described voice window data, obtain the voice data after preprocessing;

[0073] 5) Carry out short-time Fourier transform to the speech data after the preprocessing, draw speech spectrogram;

[0074] 6) Extracting speech feature data to obtain speech emotion features; ...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention proposes a bimodal fusion emotion recognition method based on a biological radar and voice information. Physiological information and the voice information are obtained through an external device radar and a microphone, the physiological information comprises breathing information and heartbeat information, physiological features and voice features are extracted from the physiological information and the voice information through a pre-trained feature extraction network, the physiological features and the voice features are fused, the fused features are input into a pre-trained deep convolutional neural network, the deep convolutional neural network comprises a first classifier and a second classifier with different labels, the first classifier and the second classifier are decision tree classifiers, different types of physiological emotion evaluation information and voice emotion evaluation information are respectively obtained through the first classifier and the second classifier, and the combined user emotion evaluation information is more objective and has reference value, so that the emotion recognition precision is improved.

Description

technical field [0001] This application relates to the field of emotion recognition, especially a bimodal fusion emotion recognition method based on biological radar and voice information. Background technique [0002] Emotion is a relatively complex and stable physiological evaluation and experience in physiology. It is a person's attitude experience and corresponding behavioral response to objective things. Modern people are faced with various pressures from study, life and work. Negative emotions and sub-mental health states. Emotion recognition helps people understand their own and other people's emotions, so as to adjust emotions in time, which is of great value to human mental health. Especially for some special groups, such as drivers, service personnel, medical staff, etc., the emotional health of such groups may even affect public safety and social stability. Therefore, analyzing and recognizing emotions is an important interdisciplinary research topic in the fiel...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
IPC IPC(8): A61B5/16A61B5/0205A61B5/08A61B5/024A61B5/05A61B5/00
CPCA61B5/165A61B5/4803A61B5/0205A61B5/024A61B5/05A61B5/08
Inventor 李兴广王鑫磊张继淋王笑竹宋文军臧景峰
Owner CHANGCHUN UNIV OF SCI & TECH