BERT-BiGRU-IDCNN-CRF named entity identification method based on attention mechanism

A technology of named entity recognition and attention, which is applied in the direction of neural learning methods, instruments, biological neural network models, etc., can solve problems such as labor-intensive, error-prone, and corpus-dependent, and achieve improved accuracy, fast training speed, and training The effect of fewer parameters

Pending Publication Date: 2021-04-30
CHONGQING UNIV OF POSTS & TELECOMM
View PDF3 Cites 34 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0008] 1. The first stage is to adopt a method based on dictionaries and rules. The disadvantage of this method is that it relies on the manual construction of rule templates by linguists, which is not only labor-intensive, but also has subjective factors and is prone to errors. Different fields not portable
[0009] 2. The method based on statistical machine learning still requires a lot of manual participation in feature extraction and relies heavily on corpus
[0010] 3. The disadvantage of BiLSTM-CRF, the mainstream model based on deep learning methods, is that the word vector of the word embedding layer cannot represent polysemy, which affects the recognition effect of the lower layer. In addition, when BiLSTM and BiGRU extract global context features , will ignore some local features

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • BERT-BiGRU-IDCNN-CRF named entity identification method based on attention mechanism
  • BERT-BiGRU-IDCNN-CRF named entity identification method based on attention mechanism
  • BERT-BiGRU-IDCNN-CRF named entity identification method based on attention mechanism

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0070] The technical solutions in the embodiments of the present invention will be described clearly and in detail below with reference to the drawings in the embodiments of the present invention. The described embodiments are only some of the embodiments of the invention.

[0071] The technical scheme that the present invention solves the problems of the technologies described above is:

[0072] Such as figure 1 As shown, it is a schematic flow chart of the BERT-BiGRU-IDCNN-CRF named entity recognition method based on the attention mechanism of the present invention. Include the following steps:

[0073] S1. Training the BERT model based on large-scale unlabeled predictions;

[0074] Specifically, the structure of the BERT model is as follows figure 2 As shown, it mainly includes the model embedding layer, bidirectional Transformer encoding and output layer.

[0075] The embedding layer is the input of the model, which is the sum of word embedding, position embedding, a...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention discloses a BERT-BiGRU-IDCNN-CRFnamed entity recognition method based on an attention mechanism. The method comprises the steps: training a BERT pre-training language model through large-scale label-free prediction; on the basis of the trained BERT model, constructing a complete BERT-BiGRU-IDCNN-Attention-CRF named entity recognition model; constructing an entity recognition training set, and training the complete entity recognition model on the training set; inputting an expected material to be subjected to entity recognition into the trained entity recognition model, and outputting a named entity recognition result. According to the method, feature vectors extracted by the BiGRU and IDCNN neural networks are combined, the defect that local features are ignored in the process of extracting global context features by the BiGRU neural network is overcome, and meanwhile, the attention mechanism is introduced, and weight allocation is performed on the extracted features, so that the features playing a key role in entity recognition are enhanced, irrelevant features are weakened, and the recognition effect of named entity recognition is further improved.

Description

technical field [0001] The invention belongs to the field of named entity recognition, and in particular relates to a BERT-BiGRU-IDCNN-CRF named entity recognition method based on an attention mechanism. Background technique [0002] Named entity recognition (NER) is one of the basic tasks in the field of natural language processing, which is to identify instances that embody concepts in text, that is, entities, such as names of people, places, and institutions. Named entity recognition has been widely used in tasks such as information extraction, question answering systems, intelligent translation, and knowledge graph construction. [0003] The methods for naming recognition are mainly divided into the following three categories: [0004] The first type of method is based on dictionaries and rules. This method first constructs dictionaries or rule templates artificially, and performs named entity recognition through matching. [0005] The second category is a method based...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
Patent Type & Authority Applications(China)
IPC IPC(8): G06F40/295G06N3/04G06N3/08
CPCG06F40/295G06N3/08G06N3/044G06N3/045
Inventor 张毅王爽胜何彬叶培明李克强
Owner CHONGQING UNIV OF POSTS & TELECOMM
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products