Method for deep convolutional neural network model compression

A neural network model and convolutional neural network technology, applied in the field of deep learning and artificial intelligence, can solve the problems of more parameters, the model cannot be deployed in storage space, and the network model becomes larger, achieving high compression ratio, reduced size, The effect of reducing the number of bits

Inactive Publication Date: 2018-07-24
SOUTH CHINA UNIV OF TECH +1
View PDF0 Cites 15 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

Generally speaking, in order to solve more complex computer vision problems, it is necessary to introduce more neurons or increase the number of layers of the network in the convolutional neural network, but this will inevitably lead to more parameters in the network and a larger network model
For example, the model size of the AlexNet deep convolutional neural network used to solve the classification problem of the ImageNet dataset reaches 243.9M. Obviously, a model of this size cannot be deployed on mobile terminals or embedded devices with very limited storage space.

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Method for deep convolutional neural network model compression
  • Method for deep convolutional neural network model compression
  • Method for deep convolutional neural network model compression

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0042] The following will clearly and completely describe the technical solutions in the embodiments of the present invention with reference to the accompanying drawings in the embodiments of the present invention. Obviously, the described embodiments are only some, not all, embodiments of the present invention. Based on the embodiments of the present invention, all other embodiments obtained by persons of ordinary skill in the art without creative efforts fall within the protection scope of the present invention.

[0043] The invention is mainly used to solve the model compression problem of the deep convolutional neural network. Through the five steps of removing network redundant connections based on dynamic thresholds, encoding residual connection weights, clustering weights, fine-tuning clustering results, and compressing and saving results, a set of algorithms for solving deep convolutional neural network model compression is established. Compared with the previous algor...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention discloses a method for deep convolutional neural network model compression. The method comprises the steps that a trained deep convolutional neural network model is retrained to remove redundant network connections; weights of remaining connections of various network layers of a convolutional neural network are coded; the weights of the remaining connections of the various network layers of the convolutional neural network are subjected to k-means clustering; clustering results are subjected to fine tuning; and results after fine tuning are saved, and a saved file is subjected toHuffman coding. According to the method, by setting a dynamic threshold, the connections in the network can be gently removed to enable the network to be recovered from the unfavorable condition thatthe connections are removed, and therefore the effect that the compression multiples is high under the condition of the same accuracy rate loss can be achieved; and in the coding process of the remaining connections, the bit number needed for representing an index value can be decreased by means of the used improved CSR coding method, therefore, the size of the compressed file can be decreased, and the compression ratio is increased.

Description

technical field [0001] The invention relates to the fields of deep learning and artificial intelligence, in particular to a method for compressing a deep convolutional neural network model. Background technique [0002] In recent years, deep learning algorithms have achieved a series of amazing results in the field of artificial intelligence, and deep convolutional neural networks are currently one of the most widely used and successful deep learning algorithms in the field of computer vision, a branch of artificial intelligence. . Generally speaking, in order to solve more complex computer vision problems, it is necessary to introduce more neurons or increase the number of layers of the network in the convolutional neural network, but this will inevitably lead to more parameters in the network and a larger network model. For example, the model size of the AlexNet deep convolutional neural network used to solve the classification problem of the ImageNet dataset reaches 243....

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
Patent Type & Authority Applications(China)
IPC IPC(8): H03M7/40H03M7/30G06N3/08G06K9/62
CPCH03M7/40H03M7/70G06N3/082G06N3/084G06F18/23213
Inventor 肖学锋金连文杨亚锋常天海刘汝杰孙俊
Owner SOUTH CHINA UNIV OF TECH
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products