Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Text coding representation method based on transformer model and multiple reference systems

A technology of transformer model and coding representation, applied in the field of machine understanding of natural language, can solve problems such as differences in understanding methods, and achieve the effect of solving difficult learning

Active Publication Date: 2019-11-01
深思考人工智能机器人科技(北京)有限公司
View PDF3 Cites 4 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0003] The existing pre-trained language representation method model is constructed by the word vector of words, and the human's level of understanding of text is the way of word-phrase-sentence-paragraph-chapter. The existing language model is different from the human understanding method. Differences lead to certain limitations in information granularity
[0004] In addition, in the models of existing pre-trained language representation methods, such as BERT, ELMO, ULMFiT, etc., not only the embedding layer but also the prediction layer (reference frame calculation layer) use a single word vector table as the reference frame, which makes the pre-training During the training language representation training process, different semantics of the same word converge with reference to the same target

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Text coding representation method based on transformer model and multiple reference systems
  • Text coding representation method based on transformer model and multiple reference systems
  • Text coding representation method based on transformer model and multiple reference systems

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0048] In order to make the purpose, technical means and advantages of the present application clearer, the present application will be further described in detail below in conjunction with the accompanying drawings.

[0049] In the embodiment of the present invention, the sentences in the text are used as pre-training tasks. In view of the use of a single frame of reference to train the language model, the polysemy of words will interfere with the training effect. Therefore, at least one independent semantic meaning is set for each word. Representation, so that the most appropriate semantic representation can be derived in combination with the context to accurately train the contextualized semantic representation; in order to avoid the situation that at least one independent semantic cannot converge, the traction structure (semantic correlation) formed by weighting is used to achieve: When the actual number of semantic concepts (absolute number of semantic concepts) n of any w...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses a text coding representation method based on a transformer model and multiple reference systems, and the method comprises the steps: splicing a word vector and a segmentation character vector of a sentence where the word vector is located based on a word vector and a separator vector coding result of a context text, and obtaining a spliced word vector; mapping the spliced word vector according to at least two set semantic concepts, obtaining at least two semantic concept vectors of the word vector, and, when the absolute semantic concept number of the word vector is smaller than the set semantic concept total number, wherein the semantic concept vectors of the word vector represent convergence, finally leaving p kinds of dissimilar semantic concept vectors; selecting the most suitable semantic concept vector of the word vector in the current context from the dissimilar semantic concept vectors through maximum pooling, and taking the most suitable semantic concept vector as a semantic prediction result of the word vector in the current context; and obtaining a probability vector of the word vector, and determining a word probability under a semantic concept corresponding to the word vector according to the probability vector.

Description

technical field [0001] The invention relates to the field of machine understanding of natural language, in particular, to a text encoding representation method based on a transformer model and a multi-reference system. Background technique [0002] In order to make the machine make better use of language, words are mapped to multi-dimensional space through word vectors. Since human text expressions often need to be combined with context to form complete semantics, the semantic representation of text has become a great challenge for machine understanding. [0003] The existing pre-trained language representation method model is constructed by the word vector of words, and the human's level of understanding of text is the way of word-phrase-sentence-paragraph-chapter. The existing language model is different from the human understanding method. Differences lead to certain limitations in information granularity. [0004] In addition, in the models of existing pre-trained lang...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
IPC IPC(8): G06F16/33G06F17/27G06N20/00
CPCG06F16/3344G06N20/00
Inventor 杨志明
Owner 深思考人工智能机器人科技(北京)有限公司
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products