Training method for hybrid frequency acoustic recognition model and speech recognition method

A technology for identifying models and mixing frequencies, applied in speech recognition, speech analysis, instruments, etc., can solve the problems of model robustness and generalization limitations, insufficient training data, and cumbersome recognition model update and maintenance. Robustness and generalizability, effect of suppressing influence

Active Publication Date: 2018-09-07
YUTOU TECH HANGZHOU
View PDF11 Cites 12 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0003] Although the use of a dedicated acoustic model can make the test environment and the training environment more compatible, it will also bring many disadvantages: First, the update and maintenance of the recognition model will be very cumbersome, and it is necessary to conduct special training for each dedicated acoustic model. Update and maintenance; Second, training each dedicated acoustic model separately will make the training data of each model insufficient, and the robustness and generalization of the model will also be limited

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Training method for hybrid frequency acoustic recognition model and speech recognition method
  • Training method for hybrid frequency acoustic recognition model and speech recognition method
  • Training method for hybrid frequency acoustic recognition model and speech recognition method

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0049] The present invention will be further described below in conjunction with the accompanying drawings and specific embodiments, but not as a limitation of the present invention.

[0050] Based on the above-mentioned problems existing in the prior art, a method for training a mixed-frequency acoustic recognition model is now provided. In this method, a unified mixed-frequency acoustic recognition model is formed through training, so as to separately perform the first The speech signal is acoustically recognized, and the second speech signal with a second sampling frequency is acoustically recognized. In other words, in this training method, a unified acoustic recognition model is trained for recognition on a variety of speech data with different sampling frequencies, rather than a dedicated acoustic recognition model trained for each speech data as in the traditional method.

[0051] The above training methods are as follows: figure 1 shown, including:

[0052] Step S1, ...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention discloses a training method for a hybrid frequency acoustic recognition model and a speech recognition method, and belongs to the technical field of speech recognition. The training method comprises the steps of obtaining a first type of speech features of a first speech signal, and processing the first type of speech features so as to obtain corresponding first speech training data;obtaining a first type of speech features of a second speech signal, and processing the first type of speech features so as to obtain corresponding second speech training data; obtaining a second type of speech features of the first speech signal and a second type of speech features of the second speech signal according to a power spectrum; forming a preliminary recognition model of the hybrid frequency acoustic recognition model according to pre-training of the first speech signal and the second speech signal; and performing supervised parameter training on the preliminary recognition modelaccording to the first speech training data, the second speech training data and the second type of speech features so as to form the hybrid frequency acoustic recognition model. The technical schemehas the beneficial effect that the recognition model has good robustness and generalization.

Description

technical field [0001] The invention relates to the technical field of speech recognition, in particular to a training method for a mixed-frequency acoustic recognition model and a speech recognition method. Background technique [0002] In the prior art, due to differences in usage environments, data transmission requirements, and technical means used for transmission, there are large differences between different recording devices and storage methods, and the main difference lies in the difference in sampling frequency. For example, voice data with a sampling frequency of 8kHz usually comes from telephone recordings, so in traditional voice recognition methods, the telephone data is specially used for training to form an acoustic recognition model of 8kHz data. Correspondingly, the speech data with a sampling frequency of 16kHz usually comes from desktop recordings, and the desktop data is also used to train an acoustic recognition model of 16kHz data. Therefore, the trad...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
Patent Type & Authority Applications(China)
IPC IPC(8): G10L15/02
CPCG10L15/02G10L15/063G10L25/30G10L15/142G10L15/16G10L25/21G10L25/24
Inventor 范利春
Owner YUTOU TECH HANGZHOU
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products