Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Cross-domain language model training method and device, electronic equipment and storage medium

A language model and training method technology, applied in electronic equipment and storage media, cross-domain language model training method, device field, and can solve problems such as intractability

Active Publication Date: 2020-05-12
北京明朝万达科技股份有限公司
View PDF10 Cites 0 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0003] Although the theoretical basis of the language model is relatively complete, the language model often encounters some difficult problems in practical applications, and the model's vulnerability to cross domains (brittleness across domains) is the most significant problem.

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Cross-domain language model training method and device, electronic equipment and storage medium
  • Cross-domain language model training method and device, electronic equipment and storage medium
  • Cross-domain language model training method and device, electronic equipment and storage medium

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0064] The technical solutions in the embodiments of the present application will be clearly and completely described below with reference to the accompanying drawings in the embodiments of the present application. Obviously, the described embodiments are part of the embodiments of the present application, not all of the embodiments. Based on the embodiments in the present application, all other embodiments obtained by those of ordinary skill in the art without creative work fall within the protection scope of the present application.

[0065] Before explaining the cross-domain language model training method of this application, firstly, the technical principle of N-Gram Chinese statistical language model is briefly explained:

[0066] Let S denote some meaningful sentence consisting of a sequence of words w in a particular order 1 , w 2 ,...,w n composition, where n represents the length of the sentence, if you need to predict the probability of the sentence S appearing in ...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention provides a cross-domain language model training method and device, electronic equipment and a storage medium. The method comprises the steps of obtaining corpus training sets of multipledomains; training a plurality of language models based on the corpus training sets of the plurality of fields to obtain respective outputs of the plurality of language models, the plurality of language models being in one-to-one correspondence with the corpus training sets of the plurality of fields; and according to respective outputs of the plurality of language models and preset interpolationcoefficients of the plurality of language models, interpolating the plurality of language models to obtain a cross-domain language model. According to the cross-domain language model training method provided by the invention, language models in multiple domains are mixed into one model through a language model mixing method based on linear interpolation, so that the cross-domain vulnerability of the language model is effectively solved, the performance index of the language model is improved, and the confusion degree of the language model is reduced.

Description

technical field [0001] The present invention relates to the technical field of natural language processing, in particular to a cross-domain language model training method, device, electronic equipment and storage medium. Background technique [0002] Natural language processing (NLP) is the intersection of computer science, linguistics and machine learning. It is dedicated to enabling computers to understand and generate human language. However, computers cannot directly perform mathematical operations on language characters, and need to convert text into mathematical symbols. Then apply various algorithms for processing to get the results we want. The language model is an abstract mathematical modeling of language based on the objective facts of language, and converts natural language text into mathematical symbols with semantic features. Therefore, the performance of the language model directly affects the performance of the NLP system. [0003] Although the theoretical f...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
IPC IPC(8): G06F16/33G06F16/35
CPCG06F16/3344G06F16/35
Inventor 谢福进王志海喻波魏力孙亚东
Owner 北京明朝万达科技股份有限公司
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products