Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Structured network model compression acceleration method based on multistage pruning

A network model and structured technology, applied in biological neural network models, neural learning methods, neural architectures, etc., can solve problems such as difficulty in taking into account parameter reduction, network acceleration, limited compression effect, inability to obtain compressed storage and reduced calculations, etc., to achieve Reduce the number of floating-point operations, improve operating efficiency, and reduce hardware dependence

Active Publication Date: 2019-12-27
UNIV OF ELECTRONICS SCI & TECH OF CHINA
View PDF8 Cites 23 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

Model pruning has been widely studied as an efficient and universal model compression method, but the compression effect achieved by existing pruning methods is very limited, and many parameter-level pruning algorithms cannot obtain actual compressed storage and Reduce operations, many filter-level pruning algorithms are often difficult to balance parameter reduction and actual network acceleration

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Structured network model compression acceleration method based on multistage pruning
  • Structured network model compression acceleration method based on multistage pruning
  • Structured network model compression acceleration method based on multistage pruning

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0049] In order to make the purpose, technical solution and advantages of the present invention clearer, the present invention will be further described in detail below in conjunction with the implementation methods and accompanying drawings.

[0050] see figure 1 , the specific implementation steps of the structured network model compression acceleration method based on multi-level pruning proposed by the present invention are as follows:

[0051] S1: Obtain the pre-training model, train the original network model to be processed on the training data set, and obtain a complete network model;

[0052] S2: Based on the pre-trained model, measure the sensitivity of the convolutional layer of the original network model, and obtain the sensitivity-pruning rate change curve of each convolutional layer through the control variable method;

[0053] S3: Iterative pruning between sensitivity layers, performing single-layer pruning on the current network model according to the order of...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses a structured network model compression acceleration method based on multistage pruning, and belongs to the technical field of model compression acceleration. The method comprises the following steps: obtaining a pre-training model, and training to obtain an initial complete network model; measuring the sensitivity of the convolution layers, and obtaining a sensitivity-pruning rate curve of each convolution layer through controlling variables; carrying out single-layer pruning from low to high according to a sensitivity sequence, and finely tuning and re-training a network model; selecting a sample as a verification set, and measuring the information entropy of the filter output feature map; performing iterative flexible pruning according to the size sequence of theoutput entropy, and finely tuning and re-training the network model; and hard pruning: carrying out retraining on the network model to recover the network performance, and obtaining and storing a lightweight model. According to the method, the large-scale convolutional neural network can be compressed on the premise of maintaining the original network performance, the local memory occupation of the network can be reduced, the floating point operation and the video memory occupation during operation are reduced, and the lightweight of the network is realized.

Description

technical field [0001] The invention relates to the technical field of model compression acceleration, in particular to a multi-level pruning-based structured network model compression acceleration method. Background technique [0002] Deep convolutional neural networks are widely used in related fields such as computer vision and natural language processing, and have achieved great success. As people pay more and more attention to convolutional neural networks, more and more layers and more structured Complex networks have sprung up like mushrooms after rain, applied to more and more research fields, and put forward higher requirements for the development of hardware devices. [0003] With the rapid development of deep learning, the improvement of hardware conditions is not so rapid. The development of convolutional neural networks relies on the improvement of computing power and storage space of today's computer equipment, especially the improvement of parallel computing c...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
IPC IPC(8): G06N3/04G06N3/08
CPCG06N3/08G06N3/045
Inventor 刘欣刚吴立帅钟鲁豪韩硕王文涵代成
Owner UNIV OF ELECTRONICS SCI & TECH OF CHINA
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products