Method for abstracting abstract text based on graph knowledge and theme perception

A topic and abstract technology, applied in the field of natural language processing, can solve problems such as fragmentation and difficulty in capturing semantic-level relationships, and achieve the effect of good robustness and adaptive ability.

Inactive Publication Date: 2022-03-22
HANGZHOU DIANZI UNIV
View PDF0 Cites 1 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

However, this approach relies on external tools, which may result in semantically fragmented output
Wang and Liu et al. built word-sentence document graphs, but it is difficult to capture semantic-level relationships in this way. Therefore, how to effectively build documents into summarizable graphs is also a difficult problem.

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Method for abstracting abstract text based on graph knowledge and theme perception
  • Method for abstracting abstract text based on graph knowledge and theme perception
  • Method for abstracting abstract text based on graph knowledge and theme perception

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0018] The present invention will be further described below in conjunction with drawings and embodiments.

[0019] The present invention proposes an abstract text summarization method based on graph knowledge and topic awareness. First, we encode input documents with a pretrained language model BERT to learn contextual sentence representations, while discovering latent topics using a neural topic model (NTM). Then, we construct a heterogeneous document graph consisting of sentence and topic nodes, and simultaneously use an improved graph attention network (GAT) to update its representation. Third, obtain the expression form of the sentence node and calculate the latent semantics. Finally, the latent semantics are fed into a Transformer-based decoder for decoding to generate the final result. We conduct extensive experiments on two real-world datasets CNN / DailyMail and XSum.

[0020] A model based on BERT, neural topic model and graph neural network proposed by the present ...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention discloses an abstract text abstracting method based on graph knowledge and theme perception. The invention provides a model based on BERT, a neural topic model and a graph neural network, and the model is called GTASum. At the input end of a document encoder, using BERT to obtain hidden word vectors of the document; at the input end of a topic encoder, obtaining a word-topic distribution vector of the document by using a neural topic model; inputting the two vectors into a graph neural network for training to obtain context contents fused with theme knowledge, and generating a text by using a decoder based on Transform; meanwhile, the condition normalized LN layer provided by the invention can cooperatively train a neural topic model and a decoder, and feature selection is effectively carried out. The result shows that the method has better robustness and adaptive ability.

Description

technical field [0001] The invention belongs to the technical field of natural language processing, and designs a method for generating text summaries, specifically an abstract text summarization method based on graph knowledge and topic perception, and a text summarization method based on pre-trained language models, neural topic models, and graph neural networks method. Background technique [0002] With the development of computer performance and large-scale language models, the task of natural language processing (NLP) has achieved significant development. As one of the core problems of natural language processing tasks, the summarization task aims to allow people to quickly grasp the important information in the text. Text summarization has been widely used in many fields, such as news, finance, conference and medical treatment. Currently, there are mainly two methods for summarization tasks: extractive methods and abstract methods. Extractive methods mainly copy imp...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
Patent Type & Authority Applications(China)
IPC IPC(8): G06F40/253G06F40/211G06F40/216G06F40/30G06N3/04G06N3/08
CPCG06F40/253G06F40/211G06F40/30G06F40/216G06N3/04G06N3/08G06N3/044
Inventor 姜明邹一凡张旻
Owner HANGZHOU DIANZI UNIV
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products