A Feature Extraction Method Fused with Inter-class Standard Deviation in Acoustic Scene Classification

A scene classification and feature extraction technology, applied in speech analysis, speech recognition, instruments, etc., can solve the problems of inconsistent perceptual resolution, affecting the classification and recognition rate of acoustic scenes, insufficient feature expression, etc., to achieve convenient implementation, improve Recognition performance, effect of simple system structure

Active Publication Date: 2020-07-10
WUHAN UNIV
View PDF16 Cites 0 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

The common extraction feature in audio is the mel spectrum. The mel spectrum is a spectrogram extracted based on the resolution of the human ear's perception of frequency, and the resolution of the acoustic scene of each frequency component may not be completely consistent with the perceptual resolution. Only a single feature is used. Spectrum is used as CNN feature input, and there is a problem of insufficient feature expression, which will affect the recognition rate of acoustic scene classification

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • A Feature Extraction Method Fused with Inter-class Standard Deviation in Acoustic Scene Classification
  • A Feature Extraction Method Fused with Inter-class Standard Deviation in Acoustic Scene Classification
  • A Feature Extraction Method Fused with Inter-class Standard Deviation in Acoustic Scene Classification

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0031] The present invention will be further described below in conjunction with embodiment:

[0032] The acoustic scene classification system based on the fusion of inter-class standard deviation features provided by the embodiment of the present invention specifically includes the following parts, and each module may be realized by using software solidification technology during specific implementation.

[0033] The feature generation module of the non-linear mapping of the standard deviation in the frequency domain between classes: According to the input audio, the output is a spectral image feature (Frequency Standard Deviation based SIF, FSD-SIF) representing the sound scene based on the standard deviation in the frequency domain. Spectral image feature generation method based on frequency domain standard deviation:

[0034] Step 1, use the audio in DCASE2017 as the original audio training set for reference, and record it as the original training set A;

[0035] Step 2, ...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention provides a feature extraction method integrating between-class standard deviation in sound scene classification. The feature extraction method comprises the following steps: S1, carrying out feature extraction based on a traditional mode, namely, calculating a frequency spectrum diagram of original audio, and carrying out downsampling based on a traditional filter, thus obtaining a feature frequency spectrum diagram P1 after downsampling; S2, carrying out feature extraction based on between-class standard deviation, namely, calculating a frequency spectrum diagram of original audio, and carrying out downsampling based on a between-class frequency domain standard deviation filter, thus obtaining a between-class standard deviation feature frequency spectrum diagram P2 after downsampling; and S3, carrying out feature fusion based on the between-class standard deviation, namely, splicing the feature frequency spectrum diagram P1 in S1 and the feature frequency spectrum diagram P2 in S2, wherein the splicing result is taken as input for a sound scene classification model. With the technical scheme for improving the sound scene classification accuracy rate, the problem that the existing sound scene resolution is not high is solved; the features are extracted through the between-class standard deviation for the first time and are fused with other features, and therefore, the identification property of the system is improved. The system provided by the invention is simple in structure and convenient to implement.

Description

technical field [0001] The invention relates to the field of sound signal analysis, in particular to a feature extraction method for integrating standard deviations between classes in sound scene classification. Background technique [0002] In recent years, in the field of audio research, under the attention of many scholars, the task of speech recognition has made great progress. However, non-speech sounds such as environmental sounds also contain important information, so their analysis and understanding are also important. of equal importance. The concept of Acoustic scene classification (ASC) is to identify the environment in which the voice clip was recorded by analyzing the voice clip, and assign the corresponding environmental semantic label to the audio. Such as parks, subways, offices, etc. The main research goal of ASC is to enable computers to understand the surrounding environment by analyzing sound like the human auditory system. It is a research direction re...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
Patent Type & Authority Patents(China)
IPC IPC(8): G10L25/51G10L25/24G10L25/18G10L15/06
CPCG10L15/063G10L25/18G10L25/24G10L25/51
Inventor 杨玉红胡瑞敏江玉至陆璐艾浩军涂卫平王晓晨张会玉
Owner WUHAN UNIV
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products