Unlock instant, AI-driven research and patent intelligence for your innovation.

Bert model pre-training method and system and computer equipment

A pre-training and model technology, applied in the Internet field, can solve the problems of word embedding matrix learning variation and loss, and achieve the effect of preventing the model from being legitimate and preventing loss

Pending Publication Date: 2021-03-19
恩亿科(北京)数据科技有限公司
View PDF0 Cites 1 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0005] Based on the above pre-training method, before training word2vec, directly setting the words after the top n in frequency to the same word "UNK" will lose a lot of word information, so that many words representing different information in the preprocessing process use the same word. A word is represented; on the other hand, directly using the word2vec pre-trained word embedding matrix to initialize the Bert model word embedding matrix and training with the Bert model may lead to poor learning of the already trained word embedding matrix

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Bert model pre-training method and system and computer equipment
  • Bert model pre-training method and system and computer equipment
  • Bert model pre-training method and system and computer equipment

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0064] In order to make the purpose, technical solutions and advantages of the present application clearer, the present application will be described and illustrated below in conjunction with the accompanying drawings and embodiments. It should be understood that the specific embodiments described here are only used to explain the present application, and are not intended to limit the present application. Based on the embodiments provided in the present application, all other embodiments obtained by persons of ordinary skill in the art without creative efforts shall fall within the protection scope of the present application.

[0065] Obviously, the accompanying drawings in the following description are only some examples or embodiments of the present application, and those skilled in the art can also apply the present application to other similar scenarios. In addition, it can also be understood that although such development efforts may be complex and lengthy, for those of ...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention relates to a Bert model pre-training method and system and computer equipment, and the method comprises: an original data set obtaining step for obtaining an original data set; a data set preprocessing step for performing word segmentation processing on the original data set to obtain a word segmentation data set, performing Word2Vec model training on the word segmentation data set to obtain word embedding matrixes of all vocabularies, and sorting and encoding the vocabularies according to the occurrence frequency to obtain high-frequency vocabularies, low-frequency vocabulariesand vocabulary codes; and a Bert model pre-training step for freezing word embedding matrix parameters of the Bert model, training the Bert model based on the word embedding matrixes of all the vocabularies, reducing the learning rate and inputting the vocabulary codes to train the Bert model again. Through the method, convergence of model parameters is optimized, and model oscillation is effectively prevented.

Description

technical field [0001] The present application relates to the field of Internet technology, in particular to a Bert model pre-training method, system and computer equipment. Background technique [0002] With the rise of deep learning technology, more and more pre-trained models have been applied to natural language processing tasks, which has greatly improved the effect of the model. Early natural language pre-training used word vector methods such as word2vec to map high-dimensional sparse word vectors to dense low-dimensional vectors as input for subsequent models. With the development of deep learning, some more powerful pre-training models have been proposed, and Bert can be regarded as a representative of them. With the help of Bert's powerful pre-training, the current best results have been achieved on many tasks. [0003] In addition to being applied to natural language processing tasks, the natural language pre-training model can also be applied to other sequence d...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
IPC IPC(8): G06F40/284G06F40/126G06N20/00
CPCG06F40/284G06F40/126G06N20/00
Inventor 佘璇段少毅
Owner 恩亿科(北京)数据科技有限公司