Model compression method and system based on sparse convolutional neural network, and related equipment

A technology of convolutional neural network and compression method, which is applied in biological neural network models, neural learning methods, neural architectures, etc., can solve problems such as difficult edge equipment, large model parameters and calculations, and operation, to achieve guaranteed performance, The effect of mitigating occupancy problems

Pending Publication Date: 2021-01-05
INST OF AUTOMATION CHINESE ACAD OF SCI
View PDF0 Cites 9 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0005] In order to solve the above-mentioned problems in the prior art, the present invention proposes a model compression method, system and related equipment based on sparse convolutional neural network. Under the premise of ensuring the effect of the model, the neural network model is compressed, and the model parameters are solved. It is difficult to run on resource-constrained edge devices due to the large amount of data and computation

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Model compression method and system based on sparse convolutional neural network, and related equipment
  • Model compression method and system based on sparse convolutional neural network, and related equipment
  • Model compression method and system based on sparse convolutional neural network, and related equipment

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0053] Preferred embodiments of the present invention are described below with reference to the accompanying drawings. Those skilled in the art should understand that these embodiments are only used to explain the technical principles of the present invention, and are not intended to limit the protection scope of the present invention.

[0054] figure 1 It is a schematic diagram of the main steps of the embodiment of the model compression method based on the sparse convolutional neural network of the present invention. Such as figure 1 As shown, the model compression method of this embodiment includes steps S10-S40:

[0055] In step S10, the model to be compressed is obtained by performing sparse regularization training on the model.

[0056] Specifically, a penalty factor can be added to the loss function by the method shown in formula (1):

[0057] L'=L+λR(X) (1)

[0058] Among them, L' is the loss function with the penalty factor added, L is the original loss function, λ...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention relates to the technical field of deep learning, particularly to a model compression method and system based on a sparse convolutional neural network, and related equipment, and aims toreduce resource occupation of a model. The compression method comprises the steps of performing sparse regularization training on a model to obtain a to-be-compressed model; calculating the importancescore of each filter according to the parameters of each convolution layer and the BN layer of the to-be-compressed model; setting an importance threshold according to the importance score and a preset pruning rate; cutting off the filters of which the importance scores are lower than the importance threshold and the BN layer parameters corresponding to the filters to obtain a pruned model; and finely adjusting the pruned model to ensure that the precision of the model is not lower than the preset precision. According to the method, on the premise that the model performance is not affected, large-amplitude compression of the model parameter quantity and the calculated quantity is achieved, the problem that resources are occupied by the model can be well solved, and the deep learning modelcan run on the edge computing equipment with limited resources.

Description

technical field [0001] The invention relates to the technical field of deep learning, in particular to a model compression method, system and related equipment based on a sparse convolutional neural network. Background technique [0002] Deep learning is developing rapidly and has been widely used in computer vision, speech recognition, natural language processing, automatic driving and other fields, and has great application prospects in edge devices of mobile terminals and embedded systems. [0003] As the scale of the deep neural network model continues to grow, training and running the deep learning model requires powerful computing power and a large amount of data storage space. However, the storage space and computing power of edge devices are limited, which makes the deep learning model in resource-constrained environments. Difficult to run on edge devices. Therefore, how to remove redundant parameters in the model and ensure the accuracy of the model is a problem th...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
Patent Type & Authority Applications(China)
IPC IPC(8): G06N3/08G06N3/04
CPCG06N3/082G06N3/045
Inventor 朱凤华韦越陈世超陈圆圆吕宜生熊刚叶佩军王飞跃
Owner INST OF AUTOMATION CHINESE ACAD OF SCI
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products