Check patentability & draft patents in minutes with Patsnap Eureka AI!

Method and device for discriminating pronunciation of polyphonic characters based on deep neural network

A deep neural network and deep neural network technology, applied in neural learning methods, biological neural network models, electrical digital data processing, etc., can solve the problem of low pronunciation accuracy of polyphonic characters, inaccurate prediction of uncommon pronunciations, and inapplicability. The problems of text training and pronunciation prediction of polyphonic characters can alleviate the imbalance of training samples, facilitate integration and improve accuracy.

Active Publication Date: 2021-09-17
BAIDU ONLINE NETWORK TECH (BEIJIBG) CO LTD
View PDF7 Cites 0 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0006] However, the polyphone disambiguation method adopted in the existing related art relies on the contextual word segmentation information of limited distance, and the accuracy rate of polyphonic pronunciation discrimination is not high in texts modified by long attributives or long adverbials; That is, the pronunciation that often appears in the text), the model tends to predict a common pronunciation, and the prediction of the uncommon pronunciation is inaccurate; the rules or models obtained by training are only effective for a certain polyphonic character, and multiple polyphonic characters need to be trained more A model or rule, not suitable for training and pronunciation prediction of texts containing multiple polyphonic characters

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Method and device for discriminating pronunciation of polyphonic characters based on deep neural network
  • Method and device for discriminating pronunciation of polyphonic characters based on deep neural network
  • Method and device for discriminating pronunciation of polyphonic characters based on deep neural network

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0027] Embodiments of the present application are described in detail below, examples of which are shown in the drawings, wherein the same or similar reference numerals denote the same or similar elements or elements having the same or similar functions throughout. The embodiments described below by referring to the figures are exemplary, and are intended to explain the present application, and should not be construed as limiting the present application.

[0028] figure 1 It is a flowchart of an embodiment of the method for discriminating pronunciation of polyphonic characters based on deep neural network in this application, such as figure 1 Shown, above-mentioned discriminating method based on the pronunciation of polyphonic character of deep neural network can comprise:

[0029] Step 101: Quantify and code each Chinese character in the text to be recognized, and generate feature vectors of the Chinese characters based on the quantized codes of the Chinese characters, the p...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The present application proposes a method and device for discriminating the pronunciation of polyphonic characters based on a deep neural network, wherein the method for discriminating the pronunciation of polyphonic characters based on a deep neural network includes: performing quantization coding on each Chinese character in the text to be recognized, and generating the The feature vector of the Chinese character; according to the feature vector of the Chinese character, combined with the context information of the two directions of the Chinese character, the input feature of the Chinese character is generated; the input feature of the Chinese character is respectively input into the DNN model corresponding to the initial consonant, the final syllable and the tone, and respectively obtains the first First probability, second probability and third probability; calculate the probability of various combinations of initials, finals and tones according to the first probability, second probability and third probability, and use the combination with the highest probability as the pronunciation of the Chinese character. The application can improve the accuracy of pronunciation discrimination, effectively alleviate the problem of unbalanced training samples caused by high-frequency sounds of polyphonic characters, and can solve the problem of pronunciation discrimination of multiple polyphonic characters, which is beneficial to the integration of speech synthesis systems.

Description

technical field [0001] The present application relates to the technical field of speech synthesis, and in particular to a method and device for discriminating the pronunciation of polyphonic characters based on a deep neural network. Background technique [0002] Word-to-sound conversion is an essential part of the Chinese speech synthesis system, and its accuracy directly affects the intelligibility of speech synthesis. Most Chinese characters have definite pronunciations, and the correct pronunciations can be obtained by searching a dictionary. However, there are many characters in Chinese that have two or more pronunciations. The key points and difficulties of word-sound conversion are the discrimination and disambiguation of these polyphonic characters. It is generally believed that the pronunciation of polyphonic characters is usually closely related to specific contextual information, semantics and language habits. How to automatically distinguish and analyze the pr...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Patents(China)
IPC IPC(8): G06F40/279G06N3/08
CPCG06N3/08G06F40/279
Inventor 聂志朋徐扬凯
Owner BAIDU ONLINE NETWORK TECH (BEIJIBG) CO LTD
Features
  • R&D
  • Intellectual Property
  • Life Sciences
  • Materials
  • Tech Scout
Why Patsnap Eureka
  • Unparalleled Data Quality
  • Higher Quality Content
  • 60% Fewer Hallucinations
Social media
Patsnap Eureka Blog
Learn More