End-to-end language model pre-training method and system, equipment and storage medium

A language model and pre-training technology, applied in the field of language model pre-training, can solve problems such as huge demand, low training efficiency, and increased language model deployment costs.

Pending Publication Date: 2021-04-23
PING AN TECH (SHENZHEN) CO LTD
View PDF0 Cites 5 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0008] Aiming at the huge demand for parameters and low training efficiency in the pre-training of the language model in the prior art, and at the same time increas

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • End-to-end language model pre-training method and system, equipment and storage medium
  • End-to-end language model pre-training method and system, equipment and storage medium
  • End-to-end language model pre-training method and system, equipment and storage medium

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0055] The present invention will be further described in detail below in conjunction with specific embodiments, which are explanations of the present invention rather than limitations.

[0056] In order to make the purpose, technical solutions and advantages of the embodiments of the present invention clearer, the technical solutions in the embodiments of the present invention will be clearly and completely described below in conjunction with the drawings in the embodiments of the present invention. Obviously, the described embodiments It is a part of embodiments of the present invention, but not all embodiments. Based on the embodiments of the present invention, all other embodiments obtained by persons of ordinary skill in the art without making creative efforts belong to the protection scope of the present invention.

[0057] It should be noted that, in the case of no conflict, the embodiments in the present application and the features in the embodiments can be combined w...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention provides an end-to-end language model pre-training method and system, equipment and a storage medium. The method comprises the steps: according to a preset knowledge similarity judgment rule, retrieving existing knowledge fragments similar to input knowledge fragments in knowledge from an existing knowledge base; splicing the input knowledge segment and the retrieved existing knowledge segment to obtain a spliced knowledge segment; carrying out mask processing on the spliced knowledge fragments; and taking the masked spliced knowledge segments as input of language model pre-training to perform prediction training, and completing end-to-end language model pre-training. According to the method, the preset similarity judgment rule is utilized, retrieval of similar existing knowledge segments is carried out in the existing knowledge base through retrieval, the requirement of the model for parameters during training is reduced, and therefore the language model can enhance utilization of external knowledge based on retrieval, and the language model training efficiency is improved.

Description

technical field [0001] The present invention relates to a language model pre-training method, in particular to an end-to-end language model pre-training method, system, device and storage medium. Background technique [0002] Natural Language Processing (NLP), the purpose of which is to enable computers to have human listening, speaking, reading, writing, translation, questioning, answering, searching, summarizing, dialogue and chatting capabilities, and to use knowledge and common sense for reasoning and decision-making, and Support customer service, diagnosis, legal, teaching and other scenarios. Natural language understanding is considered the jewel in the crown of AI. Once there is a breakthrough, it will greatly promote the landing of AI in many important scenarios. [0003] The pre-training model is to make natural language processing enter the stage of large-scale and reproducible large-scale industrial application from the original stage of manual parameter adjustm...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
IPC IPC(8): G06F16/33G06K9/62
CPCG06F16/3344G06F18/2411G06F18/214
Inventor 谯轶轩陈浩高鹏
Owner PING AN TECH (SHENZHEN) CO LTD
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products