Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Pre-training text abstract generation method based on neural topic memory

A pre-training, themed technology, applied in neural architecture, biological neural network model, unstructured text data retrieval, etc., can solve the problem of inability to complete inverted sentences, and achieve the effect of improving satisfaction and smooth and natural sentences

Active Publication Date: 2020-01-31
NANJING UNIV OF INFORMATION SCI & TECH
View PDF2 Cites 21 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

The three main issues include: identification and processing of redundant information in documents; summary and identification of important content and key information; readability and coherence of generated abstracts
[0005] For summary summaries, existing technologies tend to understand documents one-way, and there is no way to accurately understand inverted sentences, clauses, etc.

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Pre-training text abstract generation method based on neural topic memory
  • Pre-training text abstract generation method based on neural topic memory
  • Pre-training text abstract generation method based on neural topic memory

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0059] The present invention will be described in further detail below in conjunction with the accompanying drawings.

[0060] The present invention proposes a method for generating a pre-trained text summary based on neural topic memory, which fully utilizes the role of the pre-trained language model in the process of encoding and decoding, and can realize end-to-end training without manual features. At the same time, combined with the topic memory network to encode the latent topic representation of the document, this method can use pre-trained topics and topic vocabulary as features. This can better capture the important information of the article. Put the topic-aware coding sequence into the decoder to perform soft alignment through transformer multi-attention and output the preliminary summary sequence. Then use the BERT and LSTM layers of the bidirectional context to capture the features in depth, and fine-tune the parameters to generate a smoother and more informative ...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses a pre-training text abstract generation method based on neural topic memory, and the method comprises the steps: carrying out the coding of a sequence according to a complete input sequence through employing the context modeling capability of BERT pre-training, and completing the text embedding; encoding potential topic representations by using a topic memory network according to the output sequence representations; performing matching according to theme representation, reasoning and the coded sequence representation to form a final coded representation, and then usinga decoder for generating a preliminary output sequence; and masking each word in the output sequence and feeding into the BERT, and in combination with the input sequence, a decoder based on transformer and LSTM is used to predict the refined word of each mask position to realize fine adjustment. According to the method, the BERT layer and the LSTM layer of the bidirectional context are used for deeply capturing the features, the masked abstract is fed, the parameters are finely adjusted, and each abstract word is refined, so that the text abstract which is smoother and high in information amount is generated.

Description

technical field [0001] The invention belongs to the field of natural language processing, and in particular relates to a method for generating a pre-training text summary based on neural theme memory. Background technique [0002] Due to the rapid development of social media, the information on the Internet is increasing day by day. The huge amount of information makes information retrieval more and more difficult, and information summary plays an important role for information publishers, users and search engines. Concise text summaries can help users quickly find the information they need, and providing them to search engines can improve retrieval speed. For false title parties, text summarization can also play a very good supervisory role. Due to the huge workload of manual extraction of summaries, automatic summarization technology has been widely valued and researched. [0003] The text summarization process basically includes the text analysis process: analyze and pr...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(China)
IPC IPC(8): G06F16/34G06F40/289G06N3/04
CPCG06F16/345G06N3/044G06N3/045Y02D10/00
Inventor 马廷淮潘倩金子龙田青
Owner NANJING UNIV OF INFORMATION SCI & TECH
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products