Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

A language model pre-training method combined with word class information

A language model and pre-training technology, applied in natural language data processing, unstructured text data retrieval, text database clustering/classification, etc., can solve the problems of low accuracy and high cost, and achieve improved accuracy and excellent performance. Effect

Active Publication Date: 2022-03-08
创新工场(广州)人工智能研究有限公司
View PDF7 Cites 0 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0007] Aiming at the defects of low prediction accuracy and high cost of the existing language model, the present invention provides a language model pre-training method combined with word class information

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • A language model pre-training method combined with word class information
  • A language model pre-training method combined with word class information
  • A language model pre-training method combined with word class information

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0032] In order to make the purpose, technical solutions and advantages of the present invention more clear, the present invention will be further described in detail below in conjunction with the accompanying drawings and implementation examples. It should be understood that the specific embodiments described here are only used to explain the present invention, not to limit the present invention.

[0033] see figure 1 , the first embodiment of the present invention provides a language model pre-training method combined with word class information, which includes the following steps:

[0034] S1. Provide a pre-training model and pre-training text;

[0035] S2. Extract character strings from the pre-training text and form a vocabulary;

[0036] S3. Extracting two sentences from the pre-training text as training sentences and simultaneously dividing the training sentences into single-word sequences;

[0037] S4, matching the character string in the step S2 with the words in t...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The present invention relates to the technical field of language processing, in particular to a language model pre-training method combined with word class information, which includes the following steps: S1, providing a pre-training model and a pre-training text; S2, extracting a character string and forming a vocabulary; S3 1. Extract two sentences as the training sentence and divide the training sentence into a single-word sequence; S4, match the character string in step S2 with the word in the single-word sequence, and mark the character string matched with the word in the single-word sequence ; S5. Covering or replacing a word with a preset ratio in the word sequence, and simultaneously inputting the covered or replaced training sentence and the marked character string into the pre-training model to train and optimize the pre-training model; S6. Steps S2-S5 are repeated until the pre-training model reaches the set optimization condition to obtain an optimized pre-training model. The language model pre-training method combined with class word information and the pre-training model provided by the present invention have better performance on multiple downstream tasks.

Description

【Technical field】 [0001] The invention relates to the technical field of language processing, in particular to a language model pre-training method combined with word class information. 【Background technique】 [0002] At present, the most advanced pre-training language models are divided into two categories, namely Autoregressive Model and Autoencoding Model. GPT and GPT2 are well-performing autoregressive language models. The training goal of the autoregressive model is to correctly guess the next word based on the previous text. BERT is a representative self-encoding language model. The training goal of BERT is to correctly infer the masked or replaced words according to the context. Both language models have advantages and disadvantages. Autoregressive models can only combine the preceding text, but cannot simultaneously combine contextual content to complete specific tasks. On the other hand, the self-encoding language model can use context information at the same t...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Patents(China)
IPC IPC(8): G06F16/35G06F40/284G06F40/289
CPCG06F16/35
Inventor 白佳欣宋彦
Owner 创新工场(广州)人工智能研究有限公司
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products