A machine translation model optimization method based on transformer model

An optimization method and machine translation technology, applied in the field of evolutionary computing, can solve problems such as difficulty in designing models independently

Active Publication Date: 2022-03-01
SICHUAN UNIV
View PDF3 Cites 1 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

Although the Transformer model has achieved good results in machine translation, there are still several problems: 1. The arrangement mode of the MHA layer and the FFN layer in the Transformer model with different network layers is fixed, and existing studies have shown that Transformer's different layer arrangement modes have better performance than the base Transformer model on other natural language processing tasks
3. The number of layers and hyperparameters of the Transformer model are set by experts combined with domain knowledge. If non-professionals want to use the Transformer model to solve machine translation tasks, it is difficult to independently design a model that meets expectations

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • A machine translation model optimization method based on transformer model
  • A machine translation model optimization method based on transformer model
  • A machine translation model optimization method based on transformer model

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0061] The specific embodiments of the present invention are described below so that those skilled in the art can understand the present invention, but it should be clear that the present invention is not limited to the scope of the specific embodiments. For those of ordinary skill in the art, as long as various changes Within the spirit and scope of the present invention defined and determined by the appended claims, these changes are obvious, and all inventions and creations using the concept of the present invention are included in the protection list.

[0062] A machine translation model optimization method based on the Transformer model, such as figure 1 shown, including the following steps:

[0063] S1. Initialize the population of Transformer models with multiple different structures and parameters;

[0064] Specifically, genetic coding is to express the model as a searchable individual, laying the foundation for subsequent evolutionary search. In order to allow the T...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention discloses a machine translation model optimization method based on the Transformer model, which allows Transformer individuals to have different structures and parameters by designing variable-length codes and candidate blocks, and provides a variety of candidate models for word vector learning; and then designs cross mutation The strategy enables Transformer individuals to communicate with each other, so that the excellent structure or parameters for processing word vectors can be passed on to the next generation; then an environment selection strategy is designed to generate the next generation of Transformer individuals, eliminating models with relatively poor learning word vectors, and retaining the learned word vectors. The model with better vector effect; after that, iterative evolution search to find the Transformer model with the best word vector learning effect, which is used to finally solve the machine translation task, so that the Transformer model can better learn the word vector expression in the machine translation task, and improve the performance of the machine. Accuracy of the translation task.

Description

technical field [0001] The invention relates to the field of evolutionary computing, in particular to a method for optimizing a machine translation model based on a Transformer model. Background technique [0002] Transformer is a sequence-to-sequence proposed by Google in 2017 to solve machine translation tasks. Before Transformer was proposed, machine translation models can be divided into two categories: models based on feedback neural networks or convolution-based sequence regression Model. Most of the models based on the feedback neural network are composed of RNN or LSTM structure. The input of each layer in the model depends on the output state of the previous layer. The process requires a lot of time and computing resources; the convolution-based network model is composed of a multi-layer convolutional neural network, and the number of convolution operations in this model will increase rapidly when calculating the relationship of long-distance information, such as i...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
Patent Type & Authority Patents(China)
IPC IPC(8): G06F40/58G06F40/284G06F40/253G06F40/30G06N3/00
CPCG06F40/58G06F40/284G06F40/253G06F40/30G06N3/006
Inventor 孙亚楠冯犇吴杰李思毅
Owner SICHUAN UNIV
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products