Traffic scene sound classification method and system based on deep learning

A deep learning, traffic scene technology, applied in the field of sound classification based on deep learning traffic scene, can solve the problem that the sound does not have the ability to classify the sound, the capture system misses the false alarm, and the misjudgment, etc., to improve the accuracy and capture rate. , the effect of reducing false positives and high accuracy

Pending Publication Date: 2021-09-07
SHANGHAI KEYGO TECH CO LTD
View PDF11 Cites 0 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

For the current illegal whistle capture system of motor vehicles, although the system combines sound monitoring and video monitoring, it still uses traditional methods in sound classification, and usually only deals with one type of sound, and does not have sound

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Traffic scene sound classification method and system based on deep learning
  • Traffic scene sound classification method and system based on deep learning
  • Traffic scene sound classification method and system based on deep learning

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0026] Such as figure 1 As shown, this embodiment relates to a traffic scene sound classification method based on deep learning, in which the sound signal is acquired in real time from the sound collection device, the sound signal is screened, the sound signal is normalized and the short-time Fourier transform is calculated Get the sound features, use the pre-trained deep learning model to analyze the sound features, the deep learning model will output the predicted classification matrix of the sound features, use the objective function to convert the predicted classification matrix into a confidence matrix, and then determine the All sound event categories and time intervals in which sound events occur in a sound signal, the specific steps are as follows:

[0027] Step 1) Screen the real-time sound signal obtained from the sound collection device, use the wavelet packet to decompose the sound signal, use the db4 wavelet as the analysis wavelet, and calculate the proportion of...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention discloses a traffic scene sound classification method and system based on a deep learning technology. The method comprises the steps: extracting a signal time-frequency feature from a sound signal after obtaining and screening the sound signal in real time through a sound collection device, and carrying out the analysis through a trained deep learning model, and obtaining a prediction classification matrix; and converting to obtain a confidence coefficient matrix comprising all sound event types appearing in the sound signal and corresponding time intervals of the sound event types. According to the invention, the sound signals are analyzed by using the deep learning training model, so that the types of the sound events and the occurrence time of the sound events can be accurately obtained for analysis, the accuracy and the capture rate of a monitoring snapshot system related to sound discrimination are remarkably improved, and the method has higher accuracy in classifying the sound of the traffic scene, has the capability of classifying multiple types of sound, and effectively reduces the false alarm of a monitoring and recognition system.

Description

technical field [0001] The present invention relates to a technology in the field of traffic sound recognition processing, in particular to a method and system for classifying traffic scene sounds based on deep learning. Background technique [0002] Existing sound classification technologies use calculated sound pressure levels, spectral features, MFCC features, etc. as the basis for sound classification. Usually, the requirements for sound signals are high, and the accuracy and recall of classification results are greatly affected by the usage scenarios. With the development of intelligent transportation technology and the continuous improvement of people's demand for traffic management, surveillance systems with video as the core are widely used in many traffic scenarios, and play a key role in traffic management, security and other applications. [0003] However, most of the existing monitoring systems are still based on video monitoring, and only some monitoring systems...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
IPC IPC(8): G10L25/03G10L25/30G10L25/51G06N3/08G06N3/04G06K9/00G08G1/017
CPCG10L25/03G10L25/30G10L25/51G06N3/08G08G1/0175G06N3/048G06N3/045
Inventor 邱国庆李宏斌张文彬刘迅
Owner SHANGHAI KEYGO TECH CO LTD
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products