Urban sound event classifying method based on N-DenseNet and high-dimensional mfcc features

A classification method and sound technology, applied in speech analysis, speech recognition, instruments, etc., can solve the problems of low accuracy, insufficient extraction of sound classification features, and insufficient model generalization ability, and achieve high accuracy and training speed. Fast, fast convergence effect

Active Publication Date: 2019-06-28
JIANGNAN UNIV
View PDF10 Cites 13 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0004] In order to solve the problem of insufficient sound classification feature extraction, insufficient generalization ability of the model, and low classification accuracy, the present invention provides a method for classifying urban sound events based on N-DenseNet and high-dimensional mfcc features. The data can provide rich and effective feature information, the model has strong generalization ability, and the classification has higher accuracy

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Urban sound event classifying method based on N-DenseNet and high-dimensional mfcc features
  • Urban sound event classifying method based on N-DenseNet and high-dimensional mfcc features
  • Urban sound event classifying method based on N-DenseNet and high-dimensional mfcc features

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0048] like Figure 1~Figure 4 Shown, the present invention is based on the urban sound event classification method of N-DenseNet, and it comprises the following steps:

[0049] S1: collect the audio data to be processed, preprocess the original audio signal, output the audio frame sequence,

[0050] Preprocessing operations include: sampling and quantization, pre-emphasis processing, and windowing;

[0051] S2: Perform time-domain and frequency-domain analysis on the audio frame sequence, extract high-dimensional Mel-frequency cepstral coefficients, and output feature vector sequences;

[0052] The extraction scheme of high-dimensional Mel-frequency cepstral coefficients includes the following four, and one of them is selected for implementation according to the actual data situation and equipment performance:

[0053] (1) 128mfcc + 23mfcc_d + 23mfcc_d_d

[0054](2) 108mfcc + 33mfcc_d + 33mfcc_d_d

[0055] (3) 128mfcc

[0056] (4) 108mfcc;

[0057] The structure of the ...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention provides an urban sound event classifying method based on an N-DenseNet and high-dimensional mfcc feature. By the urban sound event classifying method, during audio data processing, richer and more efficient feature information can be provided, ta model has stronger generalization ability, and classification has higher accuracy. The urban sound event classifying method comprises thefollowing steps: S1, acquiring audio data to be processed, preprocessing an original audio signal, and outputting an audio frame sequence; S2, performing time-domain and frequency-domain analysis on the audio frame sequence, extracting a high-dimensional Mel-frequency cepstrum coefficient, and outputting a feature vector sequence; S3, constructing an acoustic model, and training the acoustic modelto obtain a well-trained acoustic model; S4, processing the feature vector sequence output in the step S2, and then inputting into the well-trained acoustic model for classification recognition to obtain a recognition result, namely a classification result of a sound event, wherein the acoustic model is a network model constructed by combining the characteristics of an N-order Markov model on thebasis of a DenseNet model, namely the acoustic model is an N-order DenseNet model.

Description

technical field [0001] The invention relates to the technical field of sound recognition, in particular to an urban sound event classification method based on N-DenseNet and high-dimensional mfcc features. Background technique [0002] Building smart city complexes in modern society is a trend in urban development. Among them, using a huge sensor network to collect various data such as air quality, noise levels, and population activities in the target city, and analyzing the audio data to guide urban design is one of the ideas for building a smart city. Among them, the main research significance of studying the classification of urban sound events lies in noise monitoring, urban security, soundscape assessment, and multimedia retrieval. [0003] DenseNet is a convolutional neural network with dense connections. In this network, there is a direct connection between any two layers. The input of each layer of the network is the union of the outputs of all previous layers, and...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
Patent Type & Authority Applications(China)
IPC IPC(8): G10L25/24G10L25/18G10L25/45G10L15/14G10L25/30
Inventor 曹毅黄子龙张威翟明浩刘晨李巍张宏越
Owner JIANGNAN UNIV
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products