Supercharge Your Innovation With Domain-Expert AI Agents!

Text abstract generating method and system

A technology for generating systems and texts, applied in biological neural network models, special data processing applications, instruments, etc., can solve the problem of text abstraction with multiple impurity information, limited feature extraction, and inability to identify and extract features of high-level feature key information, etc. problems, to achieve the effect of enhancing key information, enhancing key information, and improving the generation effect

Active Publication Date: 2018-11-06
SOUTH CHINA NORMAL UNIVERSITY
View PDF6 Cites 13 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0004] However, when the text summary is generated by the encoding-decoding model, the cyclic neural network directly compresses and extracts the original text information. However, due to the limited feature extraction of the original text by the cyclic neural network, many high-level features such as representing the key points of the original text The characteristics of the information cannot be identified and extracted, and the non-key information cannot be effectively identified and excluded, so that the generated text summary contains more impurity information, so that the final generated summary cannot well represent the important information of the text

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Text abstract generating method and system

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0029] see figure 1 , which is a flowchart of a method for generating a text summary in an embodiment of the present invention. The method for generating a text summary includes the following steps:

[0030] Step S1: Obtain text information and segment the text information into multiple words.

[0031] In the present invention, the text information can be segmented into multiple words by means of an existing word segmenter or word segmentation tool.

[0032] Step S2: input the divided words into the word embedding model respectively, and obtain the word vector of each word.

[0033] In the present invention, if the word vector of the ith word uses x i representation, then the word vector set representing the text can be expressed as x={x 1 ,x 2 ,...,x i-1 ,x i}; where the size of the word vector in the word embedding model can be set to 200, where the vector here and other vectors involved later are a certain word or data expressed in a computer-readable language such a...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention relates to a text abstract generating method and system. The method comprises the steps of obtaining text information, and segmenting the text information into a plurality of words; inputting the segmented words into a word embedding model, thereby obtaining word vectors of the words; obtaining all characters which form each word, and sequentially inputting all the words which form the same word into a two-way recurrent neural network model, thereby obtaining joint vectors of the words; inputting the word vectors of the words and the joint vectors corresponding to the word vectors into a nonlinear model, thereby obtaining fusion features of the words; combining the fusion features of the words into a new input text, and inputting the new input text into a convolutional neuralnetwork, thereby obtaining high-level features which represent the text; and inputting the high-level features which represent the text into an encoding-decoding model, thereby obtaining an abstract.According to the method and the system, key information of the original text can be enhanced, and non-key information can be attenuated, so that the generated text abstract can better represent important information of the text.

Description

technical field [0001] The invention relates to the field of text data processing, in particular to a method and system for generating text summaries. Background technique [0002] With the explosive development of data, especially the rapid increase of text data, people have been unable to browse and understand all the texts of interest in time, but the omission of some important text data will cause a lot of organizational and application losses. Therefore, text summarization As the information that summarizes the important data of the text, it has become the focus of people's attention, and how to automatically generate summaries based on the text data has also become a hot research topic. [0003] At present, the existing methods for automatically generating text summaries mainly use the encoding-decoding model in machine learning. Specifically, the model first uses Recurrent Neural Networks (RNN) as an encoder to encode information from the original text. Compression a...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
IPC IPC(8): G06F17/27G06N3/04
CPCG06F40/284G06N3/045
Inventor 曾碧卿周才东
Owner SOUTH CHINA NORMAL UNIVERSITY
Features
  • R&D
  • Intellectual Property
  • Life Sciences
  • Materials
  • Tech Scout
Why Patsnap Eureka
  • Unparalleled Data Quality
  • Higher Quality Content
  • 60% Fewer Hallucinations
Social media
Patsnap Eureka Blog
Learn More