Model training method and device, text processing method and device, electronic equipment and medium

A model training and text processing technology, applied in the field of data processing, can solve problems such as weak performance, low prediction accuracy, and poor performance, and achieve the effect of improving performance and reducing misleading

Pending Publication Date: 2021-08-13
ZTE CORP +1
View PDF0 Cites 1 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0009] Among them, the methods based on adversarial training proposed by Keung et al. and Zhang et al. are not as simple and direct as self-training in terms of using target language unlabeled samples, and their performance is also weaker than self-training.
However, in the current self-training process, the prediction accuracy of the teacher model on unlabeled samples is low, which misleads the training of the student model. That is to say, the current cross-language model training method has poor performance.

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Model training method and device, text processing method and device, electronic equipment and medium
  • Model training method and device, text processing method and device, electronic equipment and medium
  • Model training method and device, text processing method and device, electronic equipment and medium

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0037] In order for those skilled in the art to better understand the technical solutions of the present application, the model training method, text processing method and device, electronic equipment, and media provided by the present application will be described in detail below with reference to the accompanying drawings.

[0038] Example embodiments will be described more fully hereinafter with reference to the accompanying drawings, but may be embodied in different forms and should not be construed as limited to the embodiments set forth herein. Rather, these embodiments are provided so that this application will be thorough and complete, and will fully convey the scope of the application to those skilled in the art.

[0039] In the case of no conflict, each embodiment of the present application and each feature in the embodiment can be combined with each other.

[0040] As used herein, the term "and / or" includes any and all combinations of at least one of the associated ...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention provides a model training method, a text processing method and device, electronic equipment and a medium. The model training method comprises the following steps: taking a student model corresponding to an (a-1) th round of iteration as a teacher model corresponding to an a th round of iteration; according to the first training sample corresponding to the a-th iteration and the teacher model corresponding to the a-th iteration, training the basic model corresponding to the a-th iteration to obtain a student model corresponding to the a-th iteration, wherein in the training process, a loss function is determined according to a first probability distribution vector and a second probability distribution vector, the first probability distribution vector is a probability distribution vector obtained by inputting a first training sample corresponding to the a-th iteration into a basic model corresponding to the a-th iteration, the second probability distribution vector is a probability distribution vector obtained by inputting the first training sample corresponding to the a-th round of iteration into the teacher model corresponding to the a-th round of iteration; and outputting a corresponding student model under the condition that the convergence condition is met.

Description

technical field [0001] The embodiments of the present application relate to the technical field of data processing, and in particular, to a model training method, a text processing method and device, electronic equipment, and media. Background technique [0002] Deep learning models have greatly promoted the development of various natural language processing tasks, and training these models often requires a large number of labeled samples. However, labeled samples often only exist in languages ​​with large populations such as English and Chinese, and most languages ​​in the world often have only a few or no labeled samples. Cross-language transfer is expected to solve this problem by transferring task knowledge between languages, that is, using the labeled task data of the source language to let the model learn the corresponding task of the target language. [0003] The current mainstream cross-language migration solution is to fine-tune the multilingual pre-training langua...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
Patent Type & Authority Applications(China)
IPC IPC(8): G06F40/295G06F16/35G06N3/04
CPCG06F40/295G06F16/35G06N3/044G06N3/045
Inventor 黄书剑浦通陈家骏张洋铭屠要峰高洪黄震江周祥生
Owner ZTE CORP
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products