Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Problem classification method and system based on multi-attention mechanism and storage medium

A problem classification and attention technology, applied in neural learning methods, text database clustering/classification, computer parts, etc., can solve the problem of a single convolutional neural network long and short-term memory network, no problem text extraction, no further mining problems Issues such as textual underlying topic information

Active Publication Date: 2020-01-24
HEFEI UNIV OF TECH
View PDF6 Cites 3 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0006] Second: The existing deep learning model does not extract the potential subject information of the question text, but only uses the convolutional neural network or the long-term memory network to extract the text features, and does not further mine the potential subject information of the question text; and some methods are only single The use of convolutional neural networks or long-short-term memory networks does not combine the advantages of the two. Convolutional neural networks can capture deep semantic features extracted from data, and long-short-term memory networks can model the temporal characteristics of text. Preserve contextual semantic information of text

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Problem classification method and system based on multi-attention mechanism and storage medium
  • Problem classification method and system based on multi-attention mechanism and storage medium
  • Problem classification method and system based on multi-attention mechanism and storage medium

Examples

Experimental program
Comparison scheme
Effect test

Embodiment

[0139] The following three data sets are used to test the technical effects of five problem classification models in the prior art and the method provided by the present invention, wherein the three data sets include:

[0140] 1. The data set provided by Baidu Lab, which includes 6205 pieces of data, that is, 6205 questions and corresponding answers, such as: Who is the author of the book "Basics of Mechanical Design"? The corresponding answers are: Yang Kezhen, Cheng Guangyun, Li Zhong;

[0141] 2. The public question set of the China Computer Federation (CCF) 2016 International Conference on Natural Language Processing and Chinese Computing Questions and Answers (hereinafter referred to as the data set NLPCC2016), which contains 9604 pieces of data, such as: Lu Xun's "Chaohua How many characters are there in the book Xi Shi? The corresponding answer is: 100 thousand words;

[0142] 3. The public question set of the CCF 2017 International Conference on Natural Language Proc...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The embodiment of the invention provides a problem classification method and system based on a multi-attention mechanism and a storage medium, and belongs to the technical field of problem classification. The problem classification method comprises the steps of obtaining a to-be-classified text; converting the text into a corresponding word vector sequence by adopting a word2vec model; forming a word vector matrix based on a questionnaire attention mechanism according to the word vector sequence; performing part-of-speech tagging and encoding on the text by adopting a preset tagging set to form a part-of-speech vector sequence; respectively calculating the coefficient of each vector in the word vector sequence by adopting a formula (1); standardizing each coefficient by adopting a formula(2); determining a word vector matrix of a part-of-speech attention mechanism by adopting a formula (3); performing convolution operation on the two word vector matrixes to form a combined matrix; andinputting the combined matrix into LSTM to obtain a feature matrix with time sequence features, then obtaining feature vectors by using a self-attention mechanism, and determining the category of thetext according to the feature vectors.

Description

technical field [0001] The present invention relates to the technical field of problem classification, in particular to a problem classification method, system and storage medium based on a multi-attention mechanism. Background technique [0002] In recent years, with the wide-scale popularization of the Internet, more and more people participate in the network information interaction, which promotes the rapid development of the question answering system. As an extension of the field of information retrieval, the question answering system can provide a correct and concise answer to the natural language questions raised by users to meet the information needs of users. Question answering systems generally include three parts: question classification, information retrieval, and answer extraction. Among them, question classification is undoubtedly the basic task of the question answering system. Only by correctly analyzing and classifying questions can the candidate answer space...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
IPC IPC(8): G06F16/33G06F16/332G06F16/35G06K9/62G06N3/04G06N3/08
CPCG06F16/3329G06F16/35G06F16/3344G06N3/08G06N3/048G06N3/044G06N3/045G06F18/241
Inventor 余本功朱梦迪汲浩敏王胡燕张强杨善林
Owner HEFEI UNIV OF TECH
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products