Speech classification network training method and device, computing equipment and storage medium

A classification network and training method technology, applied in speech analysis, computing, computer parts and other directions, can solve problems such as the inability to guarantee the accuracy of prediction, and the data expression ability is not stable enough, so as to improve the efficiency of sample learning and ensure the ability of feature expression. Effect

Active Publication Date: 2021-11-02
PING AN TECH (SHENZHEN) CO LTD
View PDF4 Cites 10 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0004] The embodiment of the present invention provides a speech classification network training method, device, computing equipment and storage medium to solve the problem that the traditional small-sample self-supervised meta-learning method is not stable enough for the data expression ability of the new prediction task and cannot guarantee accurate prediction sexual problems

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Speech classification network training method and device, computing equipment and storage medium
  • Speech classification network training method and device, computing equipment and storage medium
  • Speech classification network training method and device, computing equipment and storage medium

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0030] The following will clearly and completely describe the technical solutions in the embodiments of the present invention with reference to the accompanying drawings in the embodiments of the present invention. Obviously, the described embodiments are some of the embodiments of the present invention, but not all of them. Based on the embodiments of the present invention, all other embodiments obtained by persons of ordinary skill in the art without creative efforts fall within the protection scope of the present invention.

[0031] The speech classification network training method can be applied in such as figure 1 An application environment in which a computer device communicates with a server over a network. Computer equipment can be, but is not limited to, various personal computers, laptops, smartphones, tablets, and portable wearable devices. The server can be implemented as an independent server.

[0032] In one embodiment, as figure 2 As shown, a kind of speech ...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention relates to the technical field of artificial intelligence, in particular to a speech classification network training method and device, equipment and a storage medium. The speech classification network training method comprises the following steps: acquiring a small sample data set; taking the training audio samples of the same category as a training set for contrast model learning, pre-training a contrast model based on the training set, and calculating the model loss of the contrast model; iteratively training the comparison model through model loss to obtain a trained comparison model, wherein the trained comparison model comprises a target feature extractor; connecting the target feature extractor with a classifier to construct a speech classification network; and carrying out fine tuning on the speech classification network based on the small sample data set by adopting a small sample learning mode to obtain a trained speech classification network. According to the method, the stability of the data expression ability of the new task is ensured by introducing the supervised learning mode pre-training comparison model, so that the prediction accuracy of the model for the new task is ensured.

Description

technical field [0001] The invention relates to the technical field of artificial intelligence, in particular to a speech classification network training method, device, computing equipment and storage medium. Background technique [0002] At present, with the rapid development of artificial intelligence, speech recognition, as an important part of artificial intelligence, has broad application prospects in different fields. For example, speech emotion recognition has broad application prospects in the fields of human-computer interaction, emotion monitoring or personalized recommendation. The main process of speech emotion recognition includes data preprocessing, feature extraction and emotion classification. However, the current speech emotion recognition often requires a large amount of training data to ensure the accuracy of emotion classification. In practical training, a large amount of label data is difficult to obtain, so meta-learning methods are often used to solve...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
Patent Type & Authority Applications(China)
IPC IPC(8): G10L25/63G10L25/27G06K9/62
CPCG10L25/63G10L25/27G06F18/2415Y02T10/40
Inventor 司世景王健宗
Owner PING AN TECH (SHENZHEN) CO LTD
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products