Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

A dynamic decoding method and system for neural machine translation based on entropy

A technology of machine translation and decoding methods, applied in the fields of natural language processing and neural machine translation, which can solve problems such as error accumulation

Active Publication Date: 2022-03-29
INST OF COMPUTING TECH CHINESE ACAD OF SCI
View PDF3 Cites 0 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0008] The purpose of the present invention is to solve the problem of error accumulation caused by the difference in context information between training and inference in the decoding process of neural machine translation models

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • A dynamic decoding method and system for neural machine translation based on entropy
  • A dynamic decoding method and system for neural machine translation based on entropy
  • A dynamic decoding method and system for neural machine translation based on entropy

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0071] The inventor analyzed the relationship between the entropy value of a sentence and the BLEU value when conducting neural machine translation technology research, and found that the average entropy value of words in a sentence with a high BLEU value is smaller than the average entropy value of words in a sentence with a low BLEU value , and the BLEU value of the sentence with low entropy value is higher than the BLEU value of the sentence with high entropy value. The inventor finds that there is a correlation between the entropy value of the sentence and the BLEU value by calculating the Pearson coefficient. Therefore, the present invention proposes that in each time step of the decoding phase of the training process, not only must a certain probability be sampled to select real words or predicted words to obtain context information, but also to calculate the entropy value according to the prediction result of the previous time step, and then according to the entropy The...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The present invention proposes an entropy-based neural machine translation dynamic decoding method and system. By analyzing the relationship between the entropy value of a sentence and the BLEU value, it is found that the average entropy value of words in a sentence with a high BLEU value is lower than that in a sentence with a lower BLEU value. The average entropy value of words is small, and the BLEU value of sentences with low entropy value is generally higher than that of sentences with high entropy value. By calculating the Pearson coefficient between the entropy value of the sentence and the BLEU value, it is found that there is a correlation between the two. Therefore, the present invention proposes that in each time step of the decoding phase of the training process, not only must a certain probability be sampled to select real words or predicted words to obtain context information, but also to calculate the entropy value according to the prediction result of the previous time step, and then according to the entropy The value dynamically adjusts the weight of contextual information. Addresses the issue of error accumulation in neural machine translation models during decoding due to differences in contextual information between training and inference.

Description

technical field [0001] The present invention relates to the technical fields of natural language processing and neural machine translation, and in particular to an entropy-based dynamic decoding method and system for neural machine translation. Background technique [0002] Machine translation is an important task in natural language processing. In recent years, with the rise of deep neural networks, machine translation methods based on neural networks have made great progress and gradually become mainstream machine translation methods. The neural machine translation model mainly consists of three parts: encoder network, decoder network and attention network. [0003] The encoder network is responsible for encoding the source language sentences into a list of hidden vectors, and each word corresponds to a hidden vector representation. The encoder network is usually a multi-layer bidirectional RNN structure, where the forward RNN Sequentially read in the sequence of source...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Patents(China)
IPC IPC(8): G06F40/58G06N3/08
CPCG06N3/08
Inventor 程学旗郭嘉丰范意兴王素
Owner INST OF COMPUTING TECH CHINESE ACAD OF SCI
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products