Unlock instant, AI-driven research and patent intelligence for your innovation.

Language model pre-training method for common sense concept enhancement

A language model and pre-training technology, applied in neural learning methods, biological neural network models, natural language data processing, etc., can solve problems such as the inability to display and model common sense concept information, achieve enhanced common sense understanding, and improve the accuracy of questions and answers Effect

Pending Publication Date: 2022-03-25
中国人民解放军军事科学院军事科学信息研究中心
View PDF0 Cites 0 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

The purpose of the present invention is to overcome the problem that the existing pre-training language model cannot display and model common sense concept information, and propose a language model pre-training method for common sense concept enhancement, through which the pre-training language model can be significantly improved common sense comprehension ability, and enhance its performance on downstream common sense question answering tasks

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Language model pre-training method for common sense concept enhancement
  • Language model pre-training method for common sense concept enhancement
  • Language model pre-training method for common sense concept enhancement

Examples

Experimental program
Comparison scheme
Effect test

Embodiment

[0076] like figure 1 As shown, the embodiment of the present invention proposes a language model pre-training method for common sense concept enhancement, which mainly includes the following steps:

[0077] Step 1) Construct an unsupervised corpus based on commonsense concepts, specifically including the following steps: First, traverse a given commonsense knowledge graph G to obtain a list C={c 1 ,...,c i ,...,c n},c i is the i-th common sense concept, n is the number of common sense concepts; then traverse the collected unlabeled corpus T={t 1 ,…, t j ,...,t m}, where t j is the jth sentence, m is the number of sentences; for each sentence t in the corpus j , with each common sense concept c in the common sense concept list C i Perform hard matching to obtain a set of all common sense concepts that appear in the sentence, as well as the start and end positions of each common sense concept in the sentence, so as to obtain a single training sample u for the sentence j...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses a common sense concept enhanced language model pre-training method, which comprises the following steps: step 1) collecting a corpus set, and constructing an unsupervised corpus set based on common sense concepts, the unsupervised corpus set comprising a plurality of sentences, each sentence comprising a plurality of common sense concepts and the position of each common sense concept in the sentence; 2) based on the unsupervised corpus set, randomly covering the common sense concept to form a training sample, inputting the training sample into a pre-established language model for training, and obtaining a pre-training language model with enhanced common sense concept by predicting the covered common sense concept; and step 3) obtaining a prediction sequence of language modeling through the pre-training language model enhanced by the common sense concept. According to the method, the common sense understanding ability of the pre-training language model is effectively enhanced, and experiments prove that the question and answer accuracy can be obviously improved by finely adjusting the common sense concept enhanced pre-training language model on the common sense question and answer task.

Description

technical field [0001] The invention relates to the technical field of natural language processing, in particular to a language model pre-training method for enhancing common sense concepts. Background technique [0002] In recent years, pre-training language model technology has achieved rapid development in the field of natural language processing, and has formed a new generation of natural language processing paradigm represented by pre-training-fine-tuning. The technique aims to perform unsupervised pre-training by using a masked language modeling (MLM) loss function on a large-scale unlabeled corpus, followed by supervised fine-tuning on a small-scale labeled dataset. Taking "Situation: I am wearing a wristwatch." as an example, after randomly covering some words, the input sequence "Situation: [MASK]am wearing [MASK] wrist watch." can be obtained, and the pre-training goal of the model is to predict The masked words are "I" and "a". At present, the method of fine-tun...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(China)
IPC IPC(8): G06F16/36G06F40/211G06N3/04G06N3/08
CPCG06F16/367G06F40/211G06N3/088G06N3/045
Inventor 胡明昊罗威罗准辰谭玉珊叶宇铭田昌海宋宇毛彬周纤
Owner 中国人民解放军军事科学院军事科学信息研究中心