Compression and acceleration method of structured network model based on multi-level pruning

A network model and structured technology, applied in biological neural network models, neural learning methods, neural architectures, etc., can solve problems such as difficulty in taking into account parameter reduction network acceleration, limited compression effect, inability to obtain compressed storage and reduced operations, etc. The effect of reducing the number of floating-point operations, improving operating efficiency, and reducing hardware dependencies

Active Publication Date: 2022-07-29
UNIV OF ELECTRONICS SCI & TECH OF CHINA
View PDF8 Cites 0 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

Model pruning has been widely studied as an efficient and universal model compression method, but the compression effect achieved by existing pruning methods is very limited, and many parameter-level pruning algorithms cannot obtain actual compressed storage and Reduce operations, many filter-level pruning algorithms are often difficult to balance parameter reduction and actual network acceleration

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Compression and acceleration method of structured network model based on multi-level pruning
  • Compression and acceleration method of structured network model based on multi-level pruning
  • Compression and acceleration method of structured network model based on multi-level pruning

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0049] In order to make the objectives, technical solutions and advantages of the present invention clearer, the present invention will be further described in detail below with reference to the embodiments and accompanying drawings.

[0050] see figure 1 , the specific implementation steps of the multi-level pruning-based structured network model compression acceleration method proposed by the present invention are as follows:

[0051] S1: Obtain a pre-training model, train the original network model to be processed on the training data set, and obtain a complete network model;

[0052] S2: Based on the pre-training model, measure the sensitivity of the convolutional layer of the original network model, and obtain the sensitivity-pruning rate curve of each convolutional layer through the control variable method;

[0053] S3: Iterative pruning between sensitivity layers, single-layer pruning is performed on the current network model according to the sensitivity order from low...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention discloses a method for compression and acceleration of a structured network model based on multi-level pruning, and belongs to the technical field of model compression and acceleration. The invention includes the following steps: obtaining a pre-training model, and training to obtain an initial complete network model; measuring the sensitivity of the convolution layer, and obtaining the sensitivity-pruning rate curve of each convolution layer through a control variable; according to the sensitivity order from low to high Perform single-layer pruning, fine-tune and retrain the network model; select samples as the validation set, measure the information entropy of the filter output feature map; perform iterative flexible pruning according to the order of output entropy, fine-tune and retrain the network model; hard pruning, correct The network model is retrained to restore the network performance, and the lightweight model is obtained and saved. The invention can compress the large-scale convolutional neural network on the premise of maintaining the original network performance, can reduce the local memory occupation of the network, reduce the floating-point operation and the display memory occupation during operation, and realize the lightweight of the network.

Description

technical field [0001] The invention relates to the technical field of model compression and acceleration, in particular to a method for compression and acceleration of structured network models based on multi-level pruning. Background technique [0002] Deep convolutional neural networks are widely used in computer vision and natural language processing and other related fields, and have achieved great success. As people pay more and more attention to convolutional neural networks, there are more and more layers and structures. Complex networks have sprung up like mushrooms after a spring rain, and they have been applied to more and more research fields, and have also put forward higher requirements for the development of hardware devices. [0003] With the rapid development of deep learning, the improvement of hardware conditions is not so rapid. The development of convolutional neural networks depends on the improvement of computing power and storage space of today's comp...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
Patent Type & Authority Patents(China)
IPC IPC(8): G06N3/04G06N3/08
CPCG06N3/08G06N3/045
Inventor 刘欣刚吴立帅钟鲁豪韩硕王文涵代成
Owner UNIV OF ELECTRONICS SCI & TECH OF CHINA
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products