Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Model training method, machine translation method and related devices and equipment

A training method and machine translation technology, applied in the computer field, can solve the problems of not paying attention to the relationship, weak learning representation ability, etc., to achieve strong learning ability and good effect.

Active Publication Date: 2019-08-23
TENCENT TECH (SHENZHEN) CO LTD
View PDF7 Cites 33 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0004] However, the current model based on the multi-head attention mechanism treats each subspace independently during the training process, and does not pay attention to the correlation between each subspace. Therefore, the learning representation ability of the current model based on the multi-head attention mechanism is still relatively weak.

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Model training method, machine translation method and related devices and equipment
  • Model training method, machine translation method and related devices and equipment
  • Model training method, machine translation method and related devices and equipment

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0052] In order to enable those skilled in the art to better understand the solution of the present application, the technical solution in the embodiment of the application will be clearly and completely described below in conjunction with the accompanying drawings in the embodiment of the application. Obviously, the described embodiment is only It is a part of the embodiments of this application, not all of them. Based on the embodiments in this application, all other embodiments obtained by persons of ordinary skill in the art without making creative efforts belong to the scope of protection of this application.

[0053]The terms "first", "second", "third", "fourth", etc. (if any) in the specification and claims of the present application and the above drawings are used to distinguish similar objects, and not necessarily Used to describe a specific sequence or sequence. It is to be understood that the data so used are interchangeable under appropriate circumstances such tha...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The embodiment of the invention discloses a neural network model training method, device and equipment and a medium. The method comprises the steps: acquiring a training sample set comprising trainingsamples and standard label vectors corresponding to the training samples; inputting the training sample into a neural network model comprising a plurality of attention networks; performing nonlineartransformation on the respective output vectors of the attention networks through the neural network model to obtain feature fusion vectors corresponding to the attention networks; and obtaining a neural network model, outputting a prediction label vector according to the feature fusion vector, and adjusting model parameters of the neural network model according to a comparison result of the prediction label vector and a standard label vector until a convergence condition is met, thereby obtaining a target neural network model. The output vectors of all the attention networks are fused in a nonlinear transformation mode, so that the output vectors of all the attention networks are fully interacted, a feature fusion feature vector with more information amount is generated, and the final output representation effect is ensured to be better.

Description

technical field [0001] The present application relates to the field of computer technology, and in particular to a neural network model training method, a machine translation method, a neural network model training device, a machine translation device, equipment, and a computer-readable storage medium. Background technique [0002] In recent years, the attention mechanism (Attention Mechanism) has been widely used in various tasks of deep learning-based natural language processing (Netural Language Processing, NLP), such as machine translation, intelligent question answering, speech recognition and other tasks. [0003] At present, the multi-headed attention mechanism is widely used. The so-called multi-headed attention mechanism refers to learning different features through multiple attention networks, that is, capturing relevant information on different subspaces by computing multiple times. [0004] However, the current model based on the multi-head attention mechanism tr...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
IPC IPC(8): G06F17/28G06N3/04G06N3/08G06V10/764
CPCG06N3/08G06F40/58G06N3/045G06F40/44G06F40/30G06V10/82G06V10/764Y02D10/00G06F17/16G06F18/2148
Inventor 涂兆鹏李建王星王龙跃
Owner TENCENT TECH (SHENZHEN) CO LTD
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products