Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Neural machine translation system training acceleration method based on stacking algorithm

A machine translation and algorithm technology, applied in the field of neural machine translation, can solve the problems of high equipment requirements, slow convergence speed, long training time, etc., to achieve the effect of stable training process, enhanced robustness, and improved performance

Active Publication Date: 2020-05-19
沈阳雅译网络技术有限公司
View PDF6 Cites 12 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0010] In view of the shortcomings of existing deep neural machine translation system training, such as high equipment requirements, long training time and slow convergence speed, the technical problem to be solved by the present invention is to provide a training acceleration method for neural machine translation system based on stacking algorithm

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Neural machine translation system training acceleration method based on stacking algorithm
  • Neural machine translation system training acceleration method based on stacking algorithm
  • Neural machine translation system training acceleration method based on stacking algorithm

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0040] A training acceleration method of a deep neural machine translation system based on a stacking algorithm of the present invention comprises the following steps:

[0041] 1) A Transformer model based on the self-attention mechanism, constructing an encoding end and a decoding end including an encoding block, and introducing a memory network to store the outputs of different blocks at the encoder end, and constructing a previous Transformer model based on dynamic linear aggregation;

[0042] 2) Segment the bilingual parallel sentence pairs composed of the source language and the target language, obtain the source language sequence and the target language sequence, and convert them into dense vectors that can be recognized by the computer;

[0043] 3) Input the sentence represented by the dense vector into the encoding end and the decoding end, and write the dense vector at the encoding end into the memory network of the previous Transformer model based on dynamic linear aggr...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses a training acceleration method of a deep neural machine translation system based on a stacking algorithm. The method comprises the following steps: constructing a coding end comprising a coding block, a decoding end and a preceding Transformer model; inputting sentences expressed by dense vectors into the coding end and the decoding end, and writing the input of the codingend into a memory network; writing the output vector into the memory network after completing the operation of each coding block, and accessing the memory network to perform linear aggregation to obtain the output of the current coding block; training a current model; copying the coding block parameters of the top layer to construct a new coding block and stacking the new coding block on the current coding end to construct a model containing two coding blocks; repeating the process to construct a neural machine translation system with a deeper coding end, and training the neural machine translation system to a target layer number until convergence; and performing translating by using the trained model. According to the method, the network with 48 coding layers can be trained, and the performance of the model is improved while 1.4 times of the speed-up ratio is obtained.

Description

technical field [0001] The invention relates to a neural machine translation technology, in particular to a training acceleration method for a neural machine translation system based on a stacking algorithm. Background technique [0002] Machine translation (English: Machine Translation, often abbreviated as MT) belongs to the category of computational linguistics, is an important branch of computational linguistics, and has very important scientific research value. It is the process of converting one language into another using a computer. At the same time, machine translation has important practical value. With the rapid development of economic globalization and the Internet, machine translation technology plays an increasingly critical role in promoting political, economic, and cultural exchanges. [0003] The development of machine translation technology has been closely following the development of computer technology, information theory, linguistics and other discipl...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(China)
IPC IPC(8): G06F40/58G06N3/04G06N3/08
CPCG06N3/084G06N3/047G06N3/045Y02D10/00
Inventor 杜权朱靖波肖桐张春良
Owner 沈阳雅译网络技术有限公司
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products