Check patentability & draft patents in minutes with Patsnap Eureka AI!

Fusion and compression method of multi-source neural machine translation model

A technology of machine translation and compression methods, applied in neural learning methods, biological neural network models, neural architectures, etc., can solve the problems of poor quality of auxiliary corpus, many parameters, dependence, etc., to achieve BLEU value improvement, high accuracy, large The effect of compression

Active Publication Date: 2020-02-11
BEIJING INSTITUTE OF TECHNOLOGYGY
View PDF11 Cites 4 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0004] The purpose of the present invention is to solve the following problems in neural machine translation in the past: 1) rely on large-scale corpus for neural machine translation, the problem that the translation quality that exists when corpus is insufficient; Considering the problem of translation quality degradation caused by the poor quality of auxiliary corpus; 3) Aiming at the problem of model storage caused by the large size and many parameters of multi-source models, a fusion and compression method of multi-source neural machine translation models is proposed, using CNN fusion Multi-source neural machine translation solves the problem of insufficient corpus, uses gating mechanism to solve the problem of poor quality auxiliary corpus, and selects different quantization and compression methods according to different matrix distances to solve the problem of model storage

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Fusion and compression method of multi-source neural machine translation model
  • Fusion and compression method of multi-source neural machine translation model
  • Fusion and compression method of multi-source neural machine translation model

Examples

Experimental program
Comparison scheme
Effect test

Embodiment 1

[0041] The present invention is applicable to the neural machine translation task under the condition that the source language resources are abundant and the target language resources are scarce. For example, Chinese, English, German and other language resources are abundant, and there are many mature translation systems. However, in the translation task from Chinese to Mongolian, the parallel corpus between the two is scarce, and it is difficult to directly train an effective translation system. Based on this, the present invention utilizes Chinese corpus and Chinese-English, Chinese-German translation system to obtain parallel English and German corpus, utilizes three encoders to encode three kinds of source languages ​​(Chinese, English and German), and the result that obtains is fused so that decoded by the decoder. Using this method allows the translation model to learn more language information and optimize the translation effect.

[0042] Here, the background of the sp...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention relates to a fusion and compression method of a multi-source neural machine translation model, and belongs to the technical field of natural language processing application. The method comprises the following steps: firstly, constructing a Transformer-based multi-source machine translation model; secondly, using a CNN to fuse the coding information of the plurality of sources; thirdly, enabling the model to autonomously select whether to use an auxiliary source language by using a gating mechanism; and finally, according to different distances among the matrixes, carrying out quantitative compression on the matrixes in different modes. Wherein the multi-source machine translation model further comprises an encoder and a decoder, and the compression of the model comprises calculation of matrix distance and quantitative compression. Aiming at the problem of overlarge storage space caused by complex structure and multiple parameters of a multi-source model, the method for quantitatively compressing the model is also explored, and higher compression ratio and higher accuracy are obtained.

Description

technical field [0001] The invention relates to a neural machine translation method, in particular to a multi-source neural machine translation model fusion and compression method, and belongs to the technical field of natural language processing applications. Background technique [0002] In recent years, with the maturity of various tasks of natural language processing, the neural machine translation method based on deep learning has developed rapidly, and has replaced traditional statistical machine translation in many fields, becoming a new mainstream method in academia and industry. . The training of the neural machine translation model is based on large-scale parallel corpus, and it has achieved good results in the translation of English-Chinese, Japanese-Chinese and other languages ​​with rich corpus resources. However, the performance of neural machine translation is poor in the case of insufficient parallel corpora, such as translation between Chinese and minority ...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
IPC IPC(8): G06F40/58G06N3/04G06N3/08
CPCG06N3/08G06N3/044G06N3/045
Inventor 鉴萍郭梦雪黄河燕
Owner BEIJING INSTITUTE OF TECHNOLOGYGY
Features
  • R&D
  • Intellectual Property
  • Life Sciences
  • Materials
  • Tech Scout
Why Patsnap Eureka
  • Unparalleled Data Quality
  • Higher Quality Content
  • 60% Fewer Hallucinations
Social media
Patsnap Eureka Blog
Learn More