Self-adapting method of DNN acoustic model based on personal identity characteristics
An acoustic model and identity feature technology, applied in the field of communication, can solve the problems of mismatch between the speaker's voice of the training data and the target speaker's voice, reducing the accuracy of DNN frame classification, and unable to make full use of speaker information, etc., to improve the system. Recognition performance, overcoming the drop in accuracy, and good adaptive performance
- Summary
- Abstract
- Description
- Claims
- Application Information
AI Technical Summary
Problems solved by technology
Method used
Image
Examples
Example Embodiment
[0033] Example 1
[0034] In recent years, speaker adaptation technology has received more and more attention. Adaptive methods are divided into model domain adaptation and feature domain adaptation. The application of adaptive technology in hidden Markov-Gaussian hybrid HMM-GMM systems has been very important. Mature, but it is difficult to directly apply to the Hidden Markov-Deep Neural Network HMM-DNN system. Many research institutions have done a lot of research on deep neural network adaptation. Among these methods, the speaker adaptation method based on i-vector is very popular. However, the current adaptive research methods have not fully utilized a small amount of adaptive data, and the network structure is complex, the calculation complexity is high, and the stability is not good enough. The present invention researches and discusses the i-vector adaptive method based on deep neural network DNN, and proposes an adaptive method of DNN acoustic model based on personal id...
Example Embodiment
[0047] Example 2
[0048] The adaptive method of the DNN acoustic model based on the personal identity (i-vector) feature is the same as in embodiment 1. The extraction of the personal identity i-vector feature described in step 1 of the present invention includes the following steps
[0049] 1a) Using 39-dimensional low-dimensional feature MFCC extracted from the speech data of the test set in the open source corpus, including its first-order and second-order features, train a DNN model for non-specific speaker feature extraction;
[0050] 1b) Apply the singular value matrix decomposition technique SVD to decompose the last hidden weight matrix of the DNN model extracted from the non-specific speaker features trained in step 1a), and replace the original weight matrix with it.
[0051] 1c) Apply back propagation algorithm (BP) and gradient descent method for DNN model training, and then use the trained DNN model to extract low-dimensional features of non-specific speakers.
[0052] 1d)...
Example Embodiment
[0054] Example 3
[0055] The adaptive method of the DNN acoustic model based on the personal identity (i-vector) feature is the same as that of the embodiment 1-2. The expression of the speaker identity vector i-vector described in steps (1c) and (1d) is extracted. for:
[0056] M=m+Tx+e
[0057] Among them, M represents the GMM average super vector of a specific speaker, m represents the UBM average super vector, T represents a total feature space, x represents the extracted i-vector feature representing personal identity, and e represents the residual noise item.
[0058] This example is based on the training data in the corpus, it is easy to obtain the universal background model UBM, and the total change matrix T is obtained through the expectation maximization (EM) algorithm. The personal identity i-vcetor feature extracted by this method has good speaker distinction Sex, represents the difference information between speakers, and has the advantages of low dimensionality, fewer ...
PUM
Abstract
Description
Claims
Application Information
- R&D Engineer
- R&D Manager
- IP Professional
- Industry Leading Data Capabilities
- Powerful AI technology
- Patent DNA Extraction
Browse by: Latest US Patents, China's latest patents, Technical Efficacy Thesaurus, Application Domain, Technology Topic.
© 2024 PatSnap. All rights reserved.Legal|Privacy policy|Modern Slavery Act Transparency Statement|Sitemap