Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

A network structure optimization method based on an attention mechanism

A technology of network structure and optimization method, applied in the field of Internet and deep learning, can solve the problems of no network structure sparse, no neuron neuron layer sparse constraints, no neuron activation or sparsity, etc.

Inactive Publication Date: 2019-06-28
CHINA UNIV OF PETROLEUM (EAST CHINA)
View PDF0 Cites 11 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0005] (3) The method of randomly shielding neurons based on dropout: it can be approximated as a sparse constraint between neurons, and it is realized by shielding some neurons so that they do not work, but it has great randomness and cannot be obtained A globally effective information to constrain the activation or sparsity of neurons, and sparsity constraints are also important for neural network structures
However, the L1 and L2 norms are only for the weight matrix constrained by them, but for neurons between the same layer at a higher level, and a larger range of neural layers, there is no corresponding sparse constraint to realize the network structure. Sparse, can not better perform sparse constraints on each neuron and neural layer
The dropout method shields neurons based on randomness, which simply reduces the risk of over-fitting, but cannot reasonably and effectively reduce the number of neurons to achieve network sparsity.

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • A network structure optimization method based on an attention mechanism
  • A network structure optimization method based on an attention mechanism
  • A network structure optimization method based on an attention mechanism

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0024] The technical solutions in the embodiments of the present invention will be clearly and completely described below in conjunction with the accompanying drawings in the embodiments of the present invention. Obviously, the described embodiments are only some of the embodiments of the present invention, not all of them. Based on the embodiments of the present invention, all other embodiments obtained by persons of ordinary skill in the art without creative efforts fall within the protection scope of the present invention.

[0025] Such as Figure 4 As shown, the network structure optimization method based on the attention mechanism of the present invention is characterized by a fully connected module, a convolution module, a recurrent neural network module and a feature level module.

[0026] Combine below figure 1 , figure 2 and image 3 , the specific process of the network structure optimization method based on the attention mechanism is described in detail:

[002...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention provides a network structure optimization method based on an attention mechanism, and the method comprises the steps: carrying out the reasonable weight distribution of each module of aneural network, emphasizing or weakening the contribution of certain input data to the next processing, carrying out the design in a differentiable mode, and completing an end-to-end neural network. According to the specific method, a multi-layer neural network is used for learning a weight distribution function. The learning mode is different from a common neural network training mode, firstly, only one target network is trained and then a weight distribution network is added to the network, then parameters of the target network are fixed, the weight distribution network is trained, and the target network and the weight distribution network are trained in an iteration mode until the effect is optimal.

Description

technical field [0001] The invention relates to the field of the Internet and the field of deep learning, in particular to a network structure optimization method based on an attention mechanism. Background technique [0002] The network structure optimization method based on the attention mechanism optimizes the neural network structure by assigning reasonable weights to each module of the neural network, emphasizing or weakening the contribution of some input data to the next step of processing, and designing in a differentiable manner . The technologies closest to the present invention are: [0003] (1) Regularization method based on L1 norm: L1 regularization refers to the sum of the absolute values ​​of each element in the weight vector w. By including the L1 regularization item in the loss calculation during the iterative optimization process, the regularization can be made The matrix parameters constrained by the transformation item become smaller, in order to gener...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(China)
IPC IPC(8): G06N3/04G06N3/08
Inventor 张亚飞张卫山
Owner CHINA UNIV OF PETROLEUM (EAST CHINA)
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products