Deep neural machine translation method based on dynamic linear aggregation

A deep neural and machine translation technology, applied in the field of machine translation, can solve the problems that the network structure cannot be effectively trained, the underlying network cannot be fully trained, and the translation performance cannot be further improved, so as to achieve robust training and improve information transmission. Efficiency, strong expressive effect

Active Publication Date: 2019-05-21
沈阳雅译网络技术有限公司
View PDF4 Cites 28 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0006] In order to solve the above technical problems, the present invention provides a deep neural machine translation method based on dynamic linear aggregation, to solve the problem of neural machine translation model Transformer in the prior art in training The underlying network cannot be fully trained in the deep network, resulting in the inability to effectively train the deeper network structure, thus failing to further improve translation performance

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Deep neural machine translation method based on dynamic linear aggregation
  • Deep neural machine translation method based on dynamic linear aggregation
  • Deep neural machine translation method based on dynamic linear aggregation

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0049] Researchers believe that different layers of networks in deep networks have different capabilities for feature extraction. Therefore, when calculating the upper layer network, the calculation results of the previous layer network should be effectively used, which will be more conducive to the feature learning of the upper layer network. Based on the above theory, the present invention adds a memory network to store the network output of the intermediate layer between the same layer at the encoding end and the decoding end of the Transformer model to form a Transformer model based on dynamic linear aggregation. Using the linear multi-step method in ordinary differential equations allows the current stacked network to make full use of the results of all previous stacked calculations to improve the efficiency of information transmission, so that the network can be stacked deeper and bring positive effects on performance. In the present invention, the adopted linear multi-st...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention discloses a deep neural machine translation method based on dynamic linear aggregation. Based on a Transformer model, a memory network is added to a coding end and a decoding end of theTransformer model at the same time to store the output of a previous middle lamination. The memory network is accessed before the next laminated network calculation. A linear multi-step method based on an ordinary differential equation aggregates dense vectors stored in a memory network and semantic vectors of all previous laminates to obtain a hidden layer representation for aggregating characteristics of each layer, and the hidden layer representation obtains a semantic vector of which parameters obey standard regularization as input of a next lamination through layer regularization operation. In this way, the feature vectors extracted by all previous laminated networks are fully considered when the current laminated network is calculated, so that a deeper model is trained to improve therepresentation ability of the model, and the performance of machine translation is improved.

Description

technical field [0001] The invention belongs to the technical field of machine translation and relates to a deep neural machine translation method based on dynamic linear aggregation. Background technique [0002] Neural machine translation technology usually uses a neural network-based encoder-decoder framework to model the entire translation process end-to-end, and learns the mapping relationship between the source language and the target language through a combination of linear and nonlinear functions. The Transformer model with self-attention mechanism has achieved the best translation performance in multiple languages. Among them, the encoder is responsible for encoding the input source language sentences into dense semantic vectors, and the decoder will generate translation results corresponding to the source language through the semantic vectors. Usually, the input sentence is segmented to obtain multiple word fragments, and then the word fragments are converted into...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
Patent Type & Authority Applications(China)
IPC IPC(8): G06F17/28G06N3/08
Inventor 王强李北肖桐朱靖波
Owner 沈阳雅译网络技术有限公司
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products