Check patentability & draft patents in minutes with Patsnap Eureka AI!

Polyphone pronunciation discriminating method and device based on deep neural network

A deep neural network and deep neural network technology, applied in neural learning methods, biological neural network models, electrical digital data processing, etc., can solve uncommon pronunciation predictions that are inaccurate and unsuitable for text training and pronunciation of multiple polyphonic characters Prediction, polyphone pronunciation accuracy is not high, etc., to alleviate the imbalance of training samples, facilitate integration, and improve accuracy

Active Publication Date: 2018-02-23
BAIDU ONLINE NETWORK TECH (BEIJIBG) CO LTD
View PDF7 Cites 12 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0006] However, the polyphone disambiguation method adopted in the existing related art relies on the contextual word segmentation information of limited distance, and the accuracy rate of polyphonic pronunciation discrimination is not high in texts modified by long attributives or long adverbials; That is, the pronunciation that often appears in the text), the model tends to predict a common pronunciation, and the prediction of the uncommon pronunciation is inaccurate; the rules or models obtained by training are only effective for a certain polyphonic character, and multiple polyphonic characters need to be trained more A model or rule, not suitable for training and pronunciation prediction of texts containing multiple polyphonic characters

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Polyphone pronunciation discriminating method and device based on deep neural network
  • Polyphone pronunciation discriminating method and device based on deep neural network
  • Polyphone pronunciation discriminating method and device based on deep neural network

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0027] Embodiments of the present application are described in detail below, examples of which are shown in the drawings, wherein the same or similar reference numerals denote the same or similar elements or elements having the same or similar functions throughout. The embodiments described below by referring to the figures are exemplary, and are intended to explain the present application, and should not be construed as limiting the present application.

[0028] figure 1 It is a flowchart of an embodiment of the method for discriminating pronunciation of polyphonic characters based on deep neural network in this application, such as figure 1 Shown, above-mentioned discriminating method based on the pronunciation of polyphonic character of deep neural network can comprise:

[0029] Step 101: Quantify and code each Chinese character in the text to be recognized, and generate feature vectors of the Chinese characters based on the quantized codes of the Chinese characters, the p...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention provides a polyphone pronunciation discriminating method and device based on a deep neural network. The method includes: quantitatively coding each Chinese character in a to-be-recognized text to generate feature vectors of the Chinese characters; according to the feature vectors of the Chinese characters, combining with bidirectional context information of the Chinese characters togenerate input features of the Chinese characters; respectively inputting the input features of the Chinese characters into DNN models corresponding to initial consonant, vowel and tone to respectively acquire first probability, second probability and third probability; calculating probability of various combinations of the consonant, the vowel and the tone according to the first probability, thesecond probability and the third probability, and taking the combinations with highest probability as pronunciation of the Chinese characters. By the method, pronunciation discriminating accuracy canbe improved, the problem that training samples are unbalanced caused by high-frequency sound of polyphones is effectively relieved, the pronunciation discriminating problem of multiple polyphones canbe solved, and integration of a voice synthesis system is facilitated.

Description

technical field [0001] The present application relates to the technical field of speech synthesis, and in particular to a method and device for discriminating the pronunciation of polyphonic characters based on a deep neural network. Background technique [0002] Word-to-sound conversion is an essential part of the Chinese speech synthesis system, and its accuracy directly affects the intelligibility of speech synthesis. Most Chinese characters have definite pronunciations, and the correct pronunciations can be obtained by searching a dictionary. However, there are many characters in Chinese that have two or more pronunciations. The key points and difficulties of word-sound conversion are the discrimination and disambiguation of these polyphonic characters. It is generally believed that the pronunciation of polyphonic characters is usually closely related to specific contextual information, semantics and language habits. How to automatically distinguish and analyze the pr...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
IPC IPC(8): G06F17/27G06N3/08
CPCG06N3/08G06F40/279
Inventor 聂志朋徐扬凯
Owner BAIDU ONLINE NETWORK TECH (BEIJIBG) CO LTD
Features
  • R&D
  • Intellectual Property
  • Life Sciences
  • Materials
  • Tech Scout
Why Patsnap Eureka
  • Unparalleled Data Quality
  • Higher Quality Content
  • 60% Fewer Hallucinations
Social media
Patsnap Eureka Blog
Learn More