Multi-field neural machine translation method based on self-attention mechanism

A machine translation and attention technology, applied in the field of neural machine translation, can solve the problems of ambiguity, translation performance matching limitation, high maintenance cost, etc., and achieve the effect of improving translation effect.

Pending Publication Date: 2019-07-26
SUZHOU UNIV
View PDF2 Cites 21 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0005] The first problem is a well-known problem in domain adaptation in MT: translation performance is limited by the matching between training and test data
[0006] The second problem is more practical: in the case of model inflexibility, multi-domain translation scenarios require multiple domain-specific system architectures, and each time a new domain is involved, a dedicated domain-specific data must be used to retrain a model, lacks architectural scalability and causes high maintenance costs
Howe

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Multi-field neural machine translation method based on self-attention mechanism
  • Multi-field neural machine translation method based on self-attention mechanism
  • Multi-field neural machine translation method based on self-attention mechanism

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0050] The present invention will be further described below in conjunction with the accompanying drawings and specific embodiments, so that those skilled in the art can better understand the present invention and implement it, but the examples given are not intended to limit the present invention.

[0051] Background: Transformer model for neural machine translation based on self-attention

[0052] Transformer is a fully attention-based encoder-decoder model. Transformer has two significant advantages over attention-based recurrent neural network neural machine translation models: 1) It completely avoids recurrent networks and maximizes parallel training. 2) Using the advantages of the attention mechanism, it can obtain long-distance dependence.

[0053] 1. Encoder

[0054] In an encoder, several identical layers are stacked on top of each other. Each layer contains two sublayers, a multi-head attention layer and a position-wise feed-forward layer, which are connected with...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention discloses a multi-field neural machine translation method based on a self-attention mechanism. The invention discloses a multi-field neural machine translation method based on a self-attention mechanism. The multi-field neural machine translation method comprises the following steps: carrying out two important changes on a Transformer, wherein the first change is a self-attention mechanism based on domain perception, and the domain representation is added to a key and a value vector of the original self-attention mechanism, the weight of the attention mechanism is the degree of correlation of the query and domain aware keys, the second change is to add a domain representation learning module to learn a domain vector. The method has the beneficial effects that a domain-aware NMT model architecture is provided on the basis of a neural network architecture Transformer representing the most advanced level at present. A self-attention mechanism based on domain awareness is provided for multi-domain translation. It is known that this is a first attempt on a multi-domain NMT based on a self-attention mechanism. Meanwhile, experiments and analysis also verify that the model can significantly improve the translation effect of each field and can learn the field information of training data.

Description

technical field [0001] The invention relates to the field of neural machine translation, in particular to a multi-domain neural machine translation method based on a self-attention mechanism. Background technique [0002] With the improvement of computer computing power and the application of big data, deep learning has been further applied. Neural Machine Translation based on deep learning has attracted more and more attention. In the field of NMT, the current state-of-the-art neural network architecture Transformer is an attention-based encoder-decoder model. The main idea is to encode the sentence to be translated (collectively referred to as 'source sentence' hereinafter) into a vector representation through an encoder, and then use a decoder to decode the vector representation of the source sentence and translate it into its Corresponding translations (collectively referred to as 'target sentences' hereinafter). [0003] Neural machine translation has made significan...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
IPC IPC(8): G06F17/28G06F17/22G06F16/33
CPCG06F16/334G06F40/12G06F40/58Y02D10/00
Inventor 熊德意张诗奇
Owner SUZHOU UNIV
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products