Far-field sound classification method and device

A classification method and sound technology, applied in speech analysis, speech recognition, instruments, etc., can solve the problems of target sound interference, performance degradation, and lower accuracy of sound classification tasks, and achieve the effect of improving accuracy and good robustness

Pending Publication Date: 2020-08-14
慧言科技(天津)有限公司 +1
View PDF0 Cites 8 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

In far-field sound classification, the target sound is often disturbed by background noise and reverberation, which reduces the accuracy of the sound classification task and leads to a sharp drop in performance

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Far-field sound classification method and device
  • Far-field sound classification method and device
  • Far-field sound classification method and device

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0041] In order to make the above objects, features and advantages of the present invention more obvious and understandable, the present invention will be described in further detail below in conjunction with the accompanying drawings and specific embodiments.

[0042] One of the core concepts of the embodiments of the present invention is to provide a far-field sound classification method, including: using the self-learning ability of the artificial intelligence model to establish a far-field sound classification relationship, wherein the far-field sound established by the self-learning ability of the artificial intelligence model The classification relationship is the far-field sound classification relationship established by the convolutional neural network based on data augmentation and multi-scale information; the speech signal in the target area is obtained; the feature extraction of the speech signal is performed based on the amplitude information of the speech signal, an...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The embodiment of the invention provides a far-field sound classification method, and the method comprises the steps: building a far-field sound classification relation through the self-learning capability of an artificial intelligence model, and enabling the far-field sound classification relation built by the self-learning capability of the artificial intelligence model to be a far-field sound classification relation built based on data augmentation and a convolution neural network of multi-scale information; acquiring a voice signal in the target area; performing feature extraction on the voice signal based on the amplitude information of the voice signal to obtain a spectrogram; and inputting the spectrogram into a far-field sound classification relationship established by utilizing the self-learning ability of the artificial intelligence model to obtain a classification result. Audio data of sound classification is matched with signal distribution received by a microphone in a real environment; according to the method, noise, reverberation and other interference factors are removed, and sound classification is carried out by using a data augmentation mode, so that training data of the model can better fit data distribution of a real environment, better robustness can be obtained, and the accuracy of a sound classification task is improved.

Description

technical field [0001] The present invention relates to the technical field of sound classification, in particular to a far-field sound classification method and a far-field sound classification device. Background technique [0002] With the rapid development of the Internet and information technology, people's living standards are improving day by day, and the quality of life and work requirements are also getting higher and higher. Audio, as a medium in people's daily life and business activities, deeply affects daily life. the act of living. Audio recognition is a cutting-edge research topic in the field of pattern recognition today. As a main research branch of audio recognition, Environmental Sound Classification (ESC) has recently attracted the attention of many experts and scholars and has become a hot topic. ESC is one of the most important technologies for machines to analyze their acoustic environment, and is widely used in surveillance, smart home, scene analysis...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
Patent Type & Authority Applications(China)
IPC IPC(8): G10L15/08G10L15/16G10L15/06G10L15/20
CPCG10L15/063G10L15/08G10L15/16G10L15/20
Inventor 关昊天史昊吕永杰廖启波
Owner 慧言科技(天津)有限公司
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products