Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Neural network column sparse method based on weight saliency

A neural network and neural network model technology, applied in the direction of neural learning methods, biological neural network models, neural architectures, etc., can solve the problems of small pruning space, decreased precision, and large amount of calculation, and achieve large pruning space and less accuracy drop Effect

Inactive Publication Date: 2019-07-16
ZHEJIANG UNIV
View PDF0 Cites 8 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

As shown in Figure 3(a) and Figure 3(b), the comparison of row and column sparsification shows that compared with the row sparsity, the column sparsity has a larger pruning space, which solves the limitation of the small pruning space in the structured row sparsity strategy sexual problems
The method of the present invention solves the problem of large storage capacity and large amount of calculation of the deep learning model represented by the convolutional neural network, and the problem that the precision drops more due to the small selection space existing in other structured row sparse work Improvements on the above, and propose an accelerated compression method for structured column sparseness

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Neural network column sparse method based on weight saliency
  • Neural network column sparse method based on weight saliency
  • Neural network column sparse method based on weight saliency

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0031] In conjunction with the following implementation examples, the present invention is further described in detail. However, the neural network structured column sparse algorithm proposed by the present invention is not limited to this implementation method.

[0032] (1) Preparation work

[0033] For the neural network model to be sparse, prepare the training data set, network structure configuration file, and training process configuration file. The used data set, network structure configuration, and training process configuration are all consistent with the original training method; in ResNet-50 In the neural network structured column sparse experiment, the dataset used is ImageNet-2012, and the network structure configuration and other files used are the files used by the original model of ResNet-50 (download link: https: / / cloud6.pkuml. org / f / 06997cf3f3fc48018d61 / ).

[0034] (2) Sparse structured columns

[0035] The sparse process of the overall network model is as ...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses a neural network column sparse method based on weight saliency. When the neural network model carries out feedforward operation, the four-dimensional weight tensor can be expanded into a two-dimensional matrix in a tensor expansion mode, and therefore convolution operation is converted into matrix multiplication. Therefore, on the basis of a tensor expansion operation mode,the invention discloses a method for carrying out column sparsity on a weight matrix, namely deleting columns of the weight matrix in a neural network parameter layer to obtain a sparse model, and the dimension of the operation matrix is reduced at the moment, thereby reducing the model scale of the neural network and accelerating the operation of the neural network. Network re-training is performed on the remaining weight parameters to callback the accuracy, and re-training is stopped when the accuracy of the model is not increased any more to obtain a final model. According to the method, the deep learning model can be deployed on mobile and embedded equipment, a certain actual acceleration effect is obtained, and the application of an intelligent algorithm on a mobile terminal is promoted.

Description

technical field [0001] The invention relates to the fields of neural calculation, pattern recognition and computer vision, in particular to a method for pruning the weight of a neural network to enhance its generalization ability, reduce storage capacity and speed up operation. Background technique [0002] In recent years, Convolutional Neural Networks (CNNs) have increasingly outperformed traditional computer vision methods in classification, detection, and segmentation tasks. However, CNN also brings heavy computation and storage consumption, which hinders its deployment on mobile and embedded devices. Neural network parameter sparsity shows better performance in CNN acceleration by eliminating redundant parameters that do not contribute much to the output results. However, due to irregular sparseness caused by unstructured pruning, it is difficult to achieve acceleration on general hardware platforms. Even with sparse matrix kernels, the speedup is very limited. To so...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(China)
IPC IPC(8): G06N3/08G06N3/04G06N3/063
CPCG06N3/082G06N3/063G06N3/045
Inventor 胡浩基骆阳李翔王欢周强
Owner ZHEJIANG UNIV
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products