Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Convolutional neural network compression method, system and device based on decomposition and pruning

A technology of convolutional neural network and compression method, which is applied in the direction of biological neural network model, neural architecture, etc., can solve the problem of low compression strength and achieve the effect of improving compression strength

Inactive Publication Date: 2020-02-11
INST OF AUTOMATION CHINESE ACAD OF SCI
View PDF0 Cites 8 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0006] In order to solve the above-mentioned problems in the prior art, that is, to solve the problem that the convolutional neural network is compressed by using the low-rank approximate decomposition method or the structured sparse pruning method, resulting in a relatively small compression force, the first aspect of the present invention proposes a A convolutional neural network compression method based on decomposition and pruning, the method includes:

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Convolutional neural network compression method, system and device based on decomposition and pruning
  • Convolutional neural network compression method, system and device based on decomposition and pruning
  • Convolutional neural network compression method, system and device based on decomposition and pruning

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0037] In order to make the purpose, technical solutions and advantages of the present invention clearer, the technical solutions in the embodiments of the present invention will be clearly and completely described below in conjunction with the accompanying drawings. Obviously, the described embodiments are part of the embodiments of the present invention, rather than Full examples. Based on the embodiments of the present invention, all other embodiments obtained by persons of ordinary skill in the art without making creative efforts belong to the protection scope of the present invention.

[0038] The application will be further described in detail below in conjunction with the accompanying drawings and embodiments. It should be understood that the specific embodiments described here are only used to explain related inventions, not to limit the invention. It should also be noted that, for the convenience of description, only the parts related to the related invention are sho...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention belongs to the field of artificial intelligence, particularly relates to a convolutional neural network compression method, system and device based on decomposition and pruning, and aimsto solve the problem of relatively small compression strength caused by adopting low-rank approximate decomposition or structured sparse pruning to perform convolutional neural network compression. The method of the system comprises the following steps: adding a coefficient matrix representation layer behind each convolution layer to be compressed; carrying out sparse processing on the coefficient matrix representation layer through a low-rank approximate decomposition algorithm, and carrying out pruning processing on a filter of the corresponding convolution layer according to the sparse position of the coefficient matrix representation layer; carrying out sparse processing on the decomposed coefficient matrix representation layer by adopting a structured sparse pruning method, and carrying out pruning processing on a filter of the coefficient matrix representation layer according to the sparse position of the coefficient matrix representation layer; and training the convolutional neural network after the sparse pruning processing. According to the method, two methods of low-rank approximate decomposition and structured sparse pruning are fused, so that the defects caused by a single method are overcome, and the compression strength is improved.

Description

technical field [0001] The invention belongs to the field of artificial intelligence, and in particular relates to a convolutional neural network compression method, system and device based on decomposition and pruning. Background technique [0002] With the absolute advantages of deep learning in the field of artificial intelligence, such as computer vision, speech recognition and natural language processing, artificial intelligence research has ushered in a new round of climax. In recent years, deep learning algorithms have been gradually applied to the industry. However, due to the large parameters of deep learning models and long calculation time, the application of deep learning in mobile devices and embedded devices is limited. The neural network model compression and acceleration method can reduce model storage, reduce computing energy consumption, reduce real-time memory consumption, and shorten inference delay time. [0003] Model compression can be roughly divided...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(China)
IPC IPC(8): G06N3/04
CPCG06N3/045
Inventor 胡卫明刘雨帆阮晓峰李兵原春锋潘健
Owner INST OF AUTOMATION CHINESE ACAD OF SCI
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products