Patents
Literature
Patsnap Copilot is an intelligent assistant for R&D personnel, combined with Patent DNA, to facilitate innovative research.
Patsnap Copilot

48 results about "Sound event detection" patented technology

Method for estimating at-home activities of elderly people living alone based on sound event detection

The invention discloses a method for estimating at-home activities of elderly people living alone based on sound event detection. The method comprises the following steps that firstly, a pickup arrayis placed in the indoor space for collecting multi-channel audio data and pre-processing the audio data, wherein pre-processing comprises framing and windowing; then log Mel spectrum features are extracted from the audio data in each channel, direction of arrival (DOA) space spectrum features are extracted from the audio data in all the channels, and the log Mel spectrum features and the DOA spacespectrum features are spliced; then, the spliced features are input a convolutional neural network for feature transformation; and finally, the transformed features are input into a convolutional neural network classifier, and the type of the activities is estimated. According to the method, the spectrum features and transformed features are extracted from the multi-channel audio data, the diversity of training data can be improved, the generalization ability of the convolutional neural network classifier is effectively improved, and when the at-home activities of elderly people are estimated, the higher accuracy rate can be obtained.
Owner:SOUTH CHINA UNIV OF TECH

Sound event detection method based on full convolutional network

The invention discloses a sound event detection method based on a full convolutional neural network, and mainly solves the problems of low multi-audio event detection precision and high time complexity of an existing network. According to the implementation scheme, the method comprises the following steps of 1) performing Mel cepstrum feature extraction on an audio stream to obtain time-frequencyfeature graphs of the audio stream, and forming a training data set by using the time-frequency feature graphs; 2) establishing a full convolution multi-audio event detection network composed of a frequency convolution network, a time convolution network and a decoding convolution network from top to bottom; 3) training the full convolution multi-audio event detection network by using the data set; and 4) inputting the audio stream to be detected into the trained full convolution multi-audio event detection network for multi-audio event detection to obtain the category of the audio event and the existing starting and ending time. Simulation results show that compared with an existing network 3D-CRNN with the highest precision, the precision of the method is improved by 2%, the operation speed is improved by about 5 times, and the method can be used for safety monitoring.
Owner:XIDIAN UNIV

Lightweight abnormal sound event detection method based on adaptive width self-attention mechanism

The invention discloses a lightweight abnormal sound event detection method based on a self-adaptive width self-attention mechanism. The method comprises the following steps: firstly, performing signal processing on an audio with a label to obtain a certain time-frequency characteristic representation of the audio; secondly, the feature representation (generally a vector or a matrix) with the label is taken as input to be given to the adaptive width self-attention mechanism model, then the adaptive width self-attention mechanism model has a defined loss function and a randomly initialized attention weight, the loss value of the label is calculated according to an adaptive self-attention mechanism algorithm, and the self-adaptive width self-attention mechanism model is used as a self-adaptive width self-attention mechanism model; next, the self-adaptive attention weight is updated by using a back propagation algorithm, and updating iteration is continuously performed on three input weights of attention until a loss function reaches a minimum or ideal state; and finally, storing the weight parameter at the moment by using a lightweight method, and predicting a section of unlabeled audio by taking the weight parameter as a model, so as to quickly and accurately detect the abnormal sound event.
Owner:GUILIN UNIV OF ELECTRONIC TECH

Sound event detection method based on hole convolution recurrent neural network

The invention discloses a sound event detection method based on a cavity convolution recurrent neural network. The method comprises the following steps: extracting logarithm Mel spectrum characteristics of each sample, constructing a cavity convolution recurrent neural network, wherein the cavity convolution recurrent neural network comprises a convolution neural network, a bidirectional long-short-term memory neural network and a Sigmoid output layer, using logarithm Mel spectrum features extracted from a training sample as input to train the cavity convolution recurrent neural network, and identifying a sound event in the test sample by adopting the trained cavity convolution recurrent neural network to obtain a sound event detection result. According to the method, cavity convolution isintroduced into a convolutional neural network, and the convolutional neural network and a recurrent neural network are optimized and combined to obtain a hole convolution recurrent neural network. Compared with a traditional convolutional neural network, the void convolutional recurrent neural network has a larger receptive field under the condition that the sizes of the network parameter sets are the same, contextual information of audio samples can be more effectively utilized, and a better sound event detection result is obtained.
Owner:SOUTH CHINA UNIV OF TECH

High-recall-rate weak-annotation sound event detection method

ActiveCN112036477ASolve the problem of uneven sample distributionCharacter and pattern recognitionNeural architecturesPattern recognitionPositive sample
The invention discloses a high-recall-rate weak-annotation sound event detection method, and the method comprises the steps: setting a neural network and training data corresponding to deep learning;initializing a loss function as cross entropy loss, and adding a plurality of groups of dice losses with different weights, wherein the higher the positive sample proportion is, the larger the required weight is; training, testing and observing experimental results of only using cross entropy loss and increasing a plurality of groups of dice loss with different weights; adjusting a weight hyper-parameter in the loss, and re-performing a plurality of groups of dice loss weight values; carrying out the loop iteration to find out the best effect to complete training, and obtaining a final loss function; applying the final loss function to a neural network detection model, applying the obtained model to a sound event detection system, and obtaining packet-level prediction and frame-level prediction of a sound event through a neural network classifier. According to the method, the problem of non-uniform sample distribution caused by one-to-many classification generally adopted in sound event detection can be solved, and the F2 score paying more attention to the recall rate is effectively improved.
Owner:TSINGHUA UNIV
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products