Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Speech classification model training method and device, equipment and storage medium

A classification model and training method technology, applied in speech analysis, speech recognition, instruments, etc., can solve the problems of easy over-fitting, limited label accuracy, sample data volume and sample distribution, and high labeling costs. Reduce the effect of overfitting

Pending Publication Date: 2022-04-15
PING AN TECH (SHENZHEN) CO LTD
View PDF0 Cites 1 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

Most CNN networks mark the voice during training, and establish a connection between the voice information and the target label to train the CNN model. However, this training method is limited by the accuracy of the label, the amount of sample data, and the distribution of samples. In this case, the cost of labeling is high, and it is prone to overfitting

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Speech classification model training method and device, equipment and storage medium
  • Speech classification model training method and device, equipment and storage medium
  • Speech classification model training method and device, equipment and storage medium

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0023] The following will clearly and completely describe the technical solutions in the embodiments of the present application with reference to the drawings in the embodiments of the present application. Obviously, the described embodiments are part of the embodiments of the present application, not all of them. Based on the embodiments in this application, all other embodiments obtained by persons of ordinary skill in the art without making creative efforts belong to the scope of protection of this application.

[0024] The flow charts shown in the drawings are just illustrations, and do not necessarily include all contents and operations / steps, nor must they be performed in the order described. For example, some operations / steps can be decomposed, combined or partly combined, so the actual order of execution may be changed according to the actual situation.

[0025] It should be understood that the terms used in the specification of this application are for the purpose of ...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention relates to the field of artificial intelligence, and particularly discloses a voice classification model training method and device, equipment and a storage medium, and the method comprises the steps: obtaining sample data which comprises a sample voice and a sample label corresponding to the sample voice; preprocessing the sample voice to obtain a vector matrix corresponding to the sample voice; inputting the vector matrix into a variational information bottleneck processing network of an initial voice classification model to obtain sentence representation corresponding to the sample voice; and inputting the sentence representation into a classification network of the initial voice classification model to obtain a prediction label, and then performing iterative training on the initial voice classification model according to the sample label and the prediction label to obtain a trained voice classification model. A variational information bottleneck processing network is added into a voice classification model, so that sample information is compressed, useless information is inhibited, and an over-fitting phenomenon is reduced.

Description

technical field [0001] The present application relates to the field of artificial intelligence, in particular to a training method, device, equipment and storage medium for a speech classification model. Background technique [0002] At present, many deep learning networks (Deep Neural Networks, DNN) headed by Convolutional Neural Networks (CNN) are widely used in various downstream tasks, such as speech classification tasks. In speech classification tasks, a simple CNN can often achieve good classification results. Most CNN networks mark the voice during training, and establish a connection between the voice information and the target label to train the CNN model. However, this training method is limited by the accuracy of the label, the amount of sample data, and the distribution of samples. In this case, the labeling cost is high and overfitting is prone to occur. Contents of the invention [0003] The present application provides a training method, device, equipment ...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
IPC IPC(8): G10L15/06G10L15/08G10L15/18
CPCG10L15/06G10L15/18G10L15/08
Inventor 司世景王健宗
Owner PING AN TECH (SHENZHEN) CO LTD
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products