Music emotion classification method based on multi-modal learning

An emotion classification, multimodal technology, applied in the field of emotion recognition, can solve the problems of information overload, lack of openness of music labels, ignoring music features, etc., to avoid noise or sparseness, eliminate ambiguity and uncertainty, and unify classification standards Effect

Active Publication Date: 2020-07-28
HOHAI UNIV
View PDF1 Cites 12 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0004] With the development of Internet technology and the advancement of data storage technology, music resources have grown geometrically, and the problem of information overload has emerged. However, traditional music labels (such as genres, singers, years, etc.) lack openness and ignore music. Its own characteristics, there is a lot of room for improvement

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Music emotion classification method based on multi-modal learning
  • Music emotion classification method based on multi-modal learning
  • Music emotion classification method based on multi-modal learning

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0034] Below in conjunction with accompanying drawing and specific embodiment, further illustrate the present invention, should be understood that these embodiments are only for illustrating the present invention and are not intended to limit the scope of the present invention, after having read the present invention, those skilled in the art will understand various aspects of the present invention Modifications in equivalent forms all fall within the scope defined by the appended claims of this application.

[0035] A kind of music emotion classification method based on multimodal learning of the present invention, see figure 1 , the figure shows the algorithm flow of the embodiment of the present invention, figure 2 It is a schematic diagram of the present invention, specifically comprising the following steps:

[0036] S101, audio preprocessing: the audio data is converted from MP3 format to WAV format, and each song is divided into 5 seconds of audio with a sampling freq...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention discloses a music emotion classification method based on multi-modal learning, and the method comprises the following steps: data preprocessing: carrying out the preprocessing of the audio, lyrics and comments of music according to the needed modal information, so as to obtain the effective input of a model; representation learning: mapping each modal to a respective representation space by using different modeling modes; feature extraction: extracting feature vectors of different modals after model mapping, and reducing dimensions to the same dimension; multi-modal fusion:, carrying out cascade early fusion on the features of the three different modals, and establishing more comprehensive feature representation; and emotion classification decision making: performing supervised emotion classification on the music by using the fused features. According to the music sentiment classification method, a method based on multi-modal joint learning is provided, the defect that noise or data loss exists in a current mainstream single-modal model method can be effectively reduced, and the accuracy and stability of music sentiment classification are improved.

Description

technical field [0001] The invention relates to the fields of emotion recognition and multimodal learning, in particular to a music emotion classification method based on multimodal learning. Background technique [0002] The emotional features widely used in the field of emotion recognition include sentence-based global statistical features and speech-based temporal features, but these two types of emotional features based on different durations cannot effectively express emotional issues. At present, the most research is to use acoustic features as auxiliary semantic information for speech emotion recognition, extract the emotional information contained in speech and identify its category. Commonly used speech feature extraction methods include: fundamental frequency feature extraction, formant feature extraction, Mel frequency cepstral coefficient (MFCC) extraction, derivative-based impersonal speech emotion feature extraction, and teager energy operator-based nonlinear f...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
Patent Type & Authority Applications(China)
IPC IPC(8): G06F16/65G06F16/683G06K9/62G06N3/04G06N3/08G10L25/03G10L25/51
CPCG06F16/65G06F16/685G06N3/08G10L25/51G10L25/03G06N3/047G06N3/045G06F18/2415G06F18/25G06F18/241Y02D10/00
Inventor 李晓双韩立新李景仙
Owner HOHAI UNIV
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products