Text classification method and system based on dynamic multilayer semantic perceptron

A technology of text classification and perceptron, applied in the field of text classification method and system based on dynamic multi-layer semantic perceptron, can solve the problems of reducing the amount of model parameters, large and complex models, etc., to reduce the amount of model parameters and alleviate overfitting , the effect of reducing time complexity and space complexity

Pending Publication Date: 2022-08-02
SHANDONG INST OF BUSINESS & TECH
View PDF0 Cites 0 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0006] In order to solve the technical problems in the above-mentioned background technology, the present invention provides a text classification method and system based on a dynamic multi-layer semantic perceptron, which uses a multi-layer semantic perceptron as a feature learner, which improves the quality of feature learning and reduces the time Complexity and space complexity, the dynamic depth controller is used to independently optimize the model depth, reducing the amount of model parameters, which is suitable for solving the problem of deep learning's dependence on large-scale training samples, and the problem of too large and complex models. It provides a feasible method to broaden the practical application of deep learning

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Text classification method and system based on dynamic multilayer semantic perceptron
  • Text classification method and system based on dynamic multilayer semantic perceptron
  • Text classification method and system based on dynamic multilayer semantic perceptron

Examples

Experimental program
Comparison scheme
Effect test

Embodiment 1

[0048] This embodiment provides a text classification method based on a dynamic multi-layer semantic perceptron, which specifically includes the following steps:

[0049] Step 1. Get the training set x i Indicates the ith sample, n is the number of samples, m is the length of the word, C is the number of categories, the training set contains several texts (samples), each text corresponds to a real label y i (true category).

[0050] Among them, a text can be a sentence or an article.

[0051] In this embodiment, due to the problem of insufficient training samples in the classification of news texts, each text is a piece of news, and the category y includes "politics", "economics", "military", "sports", and "entertainment", respectively. 1 to 5 numbers, so C=5.

[0052] Step 2, using the training set to train a dynamic multi-layer semantic perceptron model (Dynamic Multi-Layer Semantics Perceptron, DMSP) to obtain a trained dynamic multi-layer semantic perceptron model. A...

Embodiment 2

[0218] This embodiment provides a text classification system based on a dynamic multi-layer semantic perceptron, which specifically includes the following modules:

[0219] A text acquisition module, which is configured to: acquire the text to be classified;

[0220] a classification module, which is configured to: adopt a dynamic multi-layer semantic perceptron model to obtain the category to which the text to be classified belongs;

[0221] The dynamic multi-layer semantic perceptron model includes a word embedding layer, a dynamic depth controller, and several layers of weighted feature learners connected in sequence, and each layer of weighted feature learners is composed of sequentially connected semantic perceptrons and base classifiers; the The output of the word embedding layer is used as the input of the semantic perceptron of all weighted feature learners, and the output of the semantic perceptron of each layer of weighted feature learners is used as the input of the...

Embodiment 3

[0228] This embodiment provides a computer-readable storage medium on which a computer program is stored, and when the program is executed by a processor, implements the text classification method based on the dynamic multi-layer semantic perceptron as described in the first embodiment above. A step of.

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention provides a text classification method and system based on a dynamic multilayer semantic perceptron. The method comprises the following steps: acquiring a to-be-classified text; adopting a dynamic multilayer semantic perceptron model to obtain a category to which the to-be-classified text belongs; wherein the dynamic multilayer semantic perceptron model comprises a word embedding layer, a plurality of layers of weighted feature learners and a dynamic depth controller, the weighted feature learners are connected in sequence, and each layer of weighted feature learner is composed of a semantic perceptron and a base classifier which are connected in sequence; the output of the word embedding layer is used as the input of the semantic perceptron of all the weighted feature learners, the output of the base classifier of each layer of weighted feature learner is input into the dynamic depth controller, and the output of the semantic perceptron of each layer of weighted feature learner is used as the input of the semantic perceptron of the next layer of weighted feature learner. The feature learning quality is improved, the time complexity and the space complexity are reduced, the model parameter quantity is reduced, and the method is suitable for the situation that training samples are insufficient.

Description

technical field [0001] The invention belongs to the technical field of natural language processing, and in particular relates to a text classification method and system based on a dynamic multi-layer semantic perceptron. Background technique [0002] The statements in this section merely provide background information related to the present invention and do not necessarily constitute prior art. [0003] In the field of Natural Language Processing (NLP), Self-Attention (SA) is a feature weight representation method used to represent the relationship between different words in the same text. Models using Multi Headed Attention (MHA) (such as Transformer, BERT, etc.) have shown excellent performance on a variety of machine learning tasks, and are widely used in text classification, machine translation, named entity recognition, image classification, etc. various fields. However, the shortcomings of the Transformer model using MHA are also obvious: (1) it relies on a large num...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
Patent Type & Authority Applications(China)
IPC IPC(8): G06F16/35G06K9/62G06F40/30G06N3/04G06N3/08G06F17/18
CPCG06F16/35G06F40/30G06N3/08G06F17/18G06N3/045G06F18/2414G06F18/2415G06F18/253G06F18/214
Inventor 唐焕玲刘孝炎王育林窦全胜鲁明羽
Owner SHANDONG INST OF BUSINESS & TECH
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products