Unlock instant, AI-driven research and patent intelligence for your innovation.

Convolutional neural network pruning method based on feature map sparsification

A convolutional neural network and feature map technology, which is applied in the field of convolutional neural network pruning based on feature map sparsification, can solve the problems of no test results and decreased accuracy

Active Publication Date: 2020-03-10
BEIJING ZHICUN WITIN TECH CORP LTD
View PDF4 Cites 19 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0004] Non-Patent Document 1 (Jaderberg M, et. al. "Speeding up convolutional neural networks with low rank expansions." Proc. arxiv, 2014.) uses the method of low-rank decomposition to split the convolution operation of the convolutional layer, achieving 4.5 times acceleration effect, but causes a 1% decrease in accuracy
[0005] Non-Patent Document 2 (Gupta S, et. al. "Deep learning with limited numerical precision." Proc. International Conference on Machine Learning, 2015.) uses 16-bit fixed-point numbers to train convolutional neural networks, and at the same time realizes the MNIST data The accuracy on the set does not decrease, but the article does not give the test results on a larger data set

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Convolutional neural network pruning method based on feature map sparsification
  • Convolutional neural network pruning method based on feature map sparsification
  • Convolutional neural network pruning method based on feature map sparsification

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0051] The specific embodiments of the present invention will be described in further detail below in conjunction with the accompanying drawings.

[0052] Select crop (tomato) disease classification as the task. Diseases include 16 types of tomato powdery mildew, early blight, spot disease, etc., and the data set is a set of crop (tomato) leaf pictures. The convolutional neural network adopts the structure of feature extraction unit superposition composed of convolutional layer + batch normalization layer + ReLu activation layer, and the final linear layer output category. The feature extraction unit is represented by C, the pooling layer is represented by M, and the linear layer Represented as L, the network structure of 16 layers is represented as [C(64), C(64), M, C(128), C(128), C(128), M, C(256), C(256) ,C(256),M,C(512),C(512),C(512),M,L], where the numbers in brackets indicate the number of channels.

[0053] like figure 1 As shown, according to the method flow chart o...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses a convolutional neural network pruning method based on feature map sparsification, and relates to how to compress a convolutional neural network to reduce the parameter quantity and calculation quantity of the convolutional neural network so as to facilitate actual deployment. In the TRAINING PROCESS, regularization is added on a feature map L1 or L2 behind an active layerin a loss function; corresponding feature map channels are enabled to have different sparseness; a convolution kernel corresponding to a corresponding channel is cut off according to the sparsity of afeature map channel under a certain pruning rate, a new accuracy is obtained by finely tuning the pruned network, the pruning rate is adjusted according to the accuracy change before and after pruning, a near-optimal pruning rate is found through multiple iterations, and maximum pruning is realized under the condition that the accuracy is not reduced. According to the method, the parameter quantity and the calculation quantity of the convolutional neural network are greatly reduced.

Description

technical field [0001] The invention relates to a convolutional neural network compression technology in the field of deep learning, in particular to a convolutional neural network pruning method based on feature map sparseness. Background technique [0002] Convolutional neural network is a common deep learning network architecture that performs well on many problems such as image classification, object detection, and image style transfer. At the same time, the operation mode of the convolutional neural network is still a black box. There is no solid theoretical basis for how to set the number of layers of the network and the number of channels of each layer for specific problems. This causes redundancy of network parameters, which in turn makes the convolutional neural network have a large amount of calculation and takes up too much memory, which is not conducive to the application of the network. Therefore, it is necessary to perform model compression of convolutional ne...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(China)
IPC IPC(8): G06N3/04G06N3/08G06K9/46G06K9/62
CPCG06N3/082G06V10/40G06N3/045G06F18/241G06N3/063G06V10/451G06V10/454G06V10/513G06V10/82G06F18/2136G06F18/2193G06N3/04G06F18/214G06F18/217
Inventor 卓成闫心刚
Owner BEIJING ZHICUN WITIN TECH CORP LTD