Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Semantic coding method of long-short-term memory network based on attention distraction

A long-term and short-term memory, distraction technology, applied in semantic analysis, neural learning methods, biological neural network models, etc., can solve the problem of not establishing a link mechanism for integrating contextual information, and improve accuracy and sentence correlation. , good integrity and fluency, the effect of improving the degree of integrity

Active Publication Date: 2021-06-25
BEIJING INSTITUTE OF TECHNOLOGYGY +2
View PDF4 Cites 4 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

Although BiLSTM improves the problems of gradient explosion and gradient disappearance, the above two problems still exist when the sequence information is too long. At the same time, although BiLSTM can obtain the bidirectional feature information of the sequence, it only connects the bidirectional output without establishing a perfect Linking mechanism for fusing contextual information

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Semantic coding method of long-short-term memory network based on attention distraction
  • Semantic coding method of long-short-term memory network based on attention distraction
  • Semantic coding method of long-short-term memory network based on attention distraction

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0067] The present invention will be described in detail below according to the accompanying drawings and examples, but the specific implementation of the present invention is not limited thereto.

[0068] This embodiment illustrates the process of applying the "Semantic Coding Method Based on Distracted Long-Short-Term Memory Network" of the present invention to the natural language generation processing scenario.

[0069] The present invention trains and tests the model in a public data set cMedQA and cMedQA1. cMedQA and cMedQA1 are a question-and-answer matching data set for Chinese medical consultation, which is widely used in some medical Chinese question-and-answer evaluations. The cMedQA data comes from medical online forums, which include 54,000 questions and corresponding 100,000 answers. cMedQA1 is an extension of cMedQA, which contains 100,000 medical questions and about 200,000 corresponding answers.

[0070] The method provided in this embodiment is a writing of...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses a semantic coding method of a long short-term memory network based on attention distraction, and belongs to the field of natural language processing and generation. Aiming at the problems of semantic deviation, gradient disappearance, gradient explosion, incomplete contextual information fusion and the like in the prior art, a neural network used by the method adds a parameter sharing unit on the basis of BiLSTM, and the capability of obtaining and fusing bidirectional feature information of a model is enhanced; an activation function in an improved deep learning model is adopted, so that the probability of occurrence of a gradient problem is reduced; for an input layer and a hidden layer, a model is constructed in an interactive space and extended LSTM mode, so that the capability of fusing context information of the model is enhanced; an attention distraction mechanism of statement structure information variables is introduced, and semantic generation is limited, so that high semantic accuracy is improved. The method is suitable for natural language generation applications such as automatic news or title writing, robot customer service, conference or diagnosis report generation and the like.

Description

technical field [0001] The invention relates to a semantic coding method based on a distracted long-short-term memory network, which belongs to the field of natural language processing generation. Background technique [0002] Natural language generation is mainly used in the fields of human-computer dialogue, abstracts, picture and video descriptions, etc. It is the core technology for applications such as automatic news or headline writing, robot customer service, meeting or diagnostic report generation. The correctness of semantics is the key to generate language. [0003] Natural language generation mostly adopts the sequence conversion form of encoding to decoding, which converts a sequence of information into another corresponding sequence of text. The process of hierarchical encoding is divided into four steps: sentence semantic encoding, discourse information encoding, decoding, and sentence probability calculation. In the text generation step, sentence semantic en...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(China)
IPC IPC(8): G06F40/258G06F40/284G06F40/30G06N3/04G06N3/08
CPCG06F40/258G06F40/30G06F40/284G06N3/08G06N3/048G06N3/045
Inventor 郭树理杨文涛韩丽娜王国威宋晓伟
Owner BEIJING INSTITUTE OF TECHNOLOGYGY
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products