Natural language processing model training method and device

A natural language processing and model technology, applied in the direction of electrical digital data processing, special data processing applications, biological neural network models, etc., can solve the problems of task model generalization loss, huge difference in data sets, information loss, etc. Effect of expressing generalization ability and representation ability, improving accuracy and generalization

Active Publication Date: 2019-08-30
ZHONGKE DINGFU BEIJING TECH DEV
View PDF5 Cites 22 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

However, the data sets of different task models are very different, so the task model can only achieve better results on this data set, and for data other than this data set, the task model needs to suffer a certain generalization loss
In addition, because the training only focuses on the aforementioned single task goal, and there is inherent implicit commonality between different texts, it will cause a certain amount of information loss

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Natural language processing model training method and device
  • Natural language processing model training method and device
  • Natural language processing model training method and device

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0027] Reference will now be made in detail to the exemplary embodiments, examples of which are illustrated in the accompanying drawings. When the following description refers to the accompanying drawings, the same numerals in different drawings refer to the same or similar elements unless otherwise indicated. The implementations described in the following exemplary examples do not represent all implementations consistent with the present invention. Rather, they are merely examples of apparatuses and methods consistent with aspects of the invention as recited in the appended claims.

[0028] The present application provides a training method and device for a natural language processing model. The "natural language processing model" mentioned in this embodiment refers to a model built based on an artificial neural network for processing natural language text, such as a classification model. Before introducing the specific implementation of the method and device of the present...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention discloses a natural language processing model training method and device. The method is applied to a plurality of natural language processing models with different task targets. A plurality of natural language processing models share a part of hidden layers, and the method comprises the following steps: fusing label information of training samples corresponding to all tasks and representing the fused label information as a multi-label matrix, calculating a hit probability according to a task target vector output by the model and the multi-label matrix, and calculating a single task loss value according to the hit probability; calculating a multi-task loss value according to each single-task loss value; and finally, adjusting parameters of each model according to the multi-task loss value. In the method, a plurality of natural language processing models are combined for learning; shallow feature representations of a plurality of tasks are shared by sharing part of hidden layers, gradients are propagated back at the same time to help the hidden layers to escape from a local minimum value, fusion of a plurality of labels and vector representations of the labels help increase the expression generalization ability and the representation ability of the labels, and then the accuracy and the generalization of each model can be improved.

Description

technical field [0001] The present application relates to the technical field of natural language processing, in particular to a training method and device for a natural language processing model. Background technique [0002] Natural language processing is an important direction in the field of computer science and artificial intelligence. It studies how to make computers understand the meaning of natural language texts and express given intentions and thoughts through natural language texts. The former is called Natural language understanding, which is called natural language generation. [0003] Whether it is natural language understanding or natural language generation, there are many tasks, which can be roughly divided into lexical analysis, sentence analysis, semantic analysis, information extraction, high-level tasks, etc. according to the task type. Among them, since all natural languages ​​have lexical and syntactic features, there is no need to limit the field of ...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
Patent Type & Authority Applications(China)
IPC IPC(8): G06F17/27G06N3/04
CPCG06F40/211G06F40/284G06F40/30G06N3/044
Inventor 李健铨刘小康马力群
Owner ZHONGKE DINGFU BEIJING TECH DEV
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products