Eureka AIR delivers breakthrough ideas for toughest innovation challenges, trusted by R&D personnel around the world.

Environmental sound recognition method based on ensemble learning and convolutional neural network

A technology of convolutional neural network and environmental sound, which is applied to biological neural network models, neural architecture, speech analysis, etc., can solve the problems of easy over-fitting and weak model generalization ability, so as to enhance generalization ability and alleviate The effect of overfitting

Pending Publication Date: 2021-01-12
江苏聆世科技有限公司
View PDF0 Cites 4 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0004] Existing recognition methods based on convolutional neural networks, recognition methods based on convolutional neural networks and recurrent neural networks, and recognition methods based on Gaussian mixture models all use existing environmental audio data to train a single model to identify unknown environmental audio. Identification, the models trained by this method have weak generalization ability and are prone to the disadvantage of overfitting

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Environmental sound recognition method based on ensemble learning and convolutional neural network
  • Environmental sound recognition method based on ensemble learning and convolutional neural network
  • Environmental sound recognition method based on ensemble learning and convolutional neural network

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0023] Such as figure 1 As shown, it is a flow chart of an environmental sound recognition method based on integrated learning and convolutional neural network, including the following steps:

[0024] S1. Feature extraction. In order to facilitate speech analysis, first gather N sampling points into one observation unit, called a frame. In order to avoid excessive changes between two adjacent frames, there will be a period of overlap between two adjacent frames. area. Substituting each frame into a window function removes possible signal discontinuities at both ends of each frame. For each short-term analysis window, the corresponding amplitude spectrum is obtained by FFT, and the energy spectrum of the sound is obtained by taking the square, and then the Mel energy spectrum of the sound is obtained by using the Mel filter bank, and then the log nonlinear transformation is performed on the Mel energy spectrum , to get the final Mel energy spectrum feature;

[0025] S2. Mode...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses an environmental sound recognition method based on ensemble learning and a convolutional neural network; and the method comprises the steps: S1, performing feature extraction:carrying out the framing and windowing of an original audio, obtaining a Mel energy spectrum of sound through a Mel filter bank, and finally obtaining a final Mel energy spectrum feature as a data set; S2, performing model training: performing model training on the data set by adopting k-fold cross validation and using a mixup data enhancement method to obtain K convolutional neural network models; and S3, performing sound testing: recognizing a sound sample to be tested through the convolutional neural network model. According to the method, k models can be trained by utilizing k-fold cross validation and k models are combined to perform voice recognition, the generalization ability of the models is greatly enhanced, and the over-fitting phenomenon is effectively relieved; in addition, for the condition of small data volume, the mixup data are used for enhancing the mixing of original samples so as to further enhance the generalization ability of the models.

Description

technical field [0001] The invention relates to the technical field of artificial intelligence, in particular to an environmental sound recognition method based on integrated learning and convolutional neural network. Background technique [0002] In the research of audio information, environmental sound recognition is an important research field, and it has great application potential in the fields of security monitoring, medical monitoring, smart home and scene analysis. Compared with speech recognition, environmental sound has characteristics such as noise-like and wide frequency spectrum, which makes the recognition of environmental sound more challenging. [0003] The existing environmental sound recognition method based on convolutional neural network usually first divides the existing data into a training set and a test set, and then uses the training set to train a model until the model converges. During the training process, the test set is used to test the model. ...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(China)
IPC IPC(8): G10L17/26G10L25/45G10L25/30G10L25/18G10L17/04G06N3/04
CPCG10L17/26G10L25/45G10L25/18G10L25/30G10L17/04G06N3/045
Inventor 陈俊谢维王震宇郭宏成
Owner 江苏聆世科技有限公司
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Eureka Blog
Learn More
PatSnap group products