Check patentability & draft patents in minutes with Patsnap Eureka AI!

Hierarchical sparse coding method of pruned deep neural network with extremely high compression ratio

A deep neural network and sparse coding technology, applied in the field of layered sparse coding, can solve the problems of difficult parallel implementation of procedural methods, difficult to achieve high compression ratio and efficient calculation, consumption and other problems at the same time, achieve considerable benefits, novel and concise layered The effect of sparse coding structure

Pending Publication Date: 2021-02-26
NANJING UNIV +1
View PDF0 Cites 6 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

Some new methods have been proposed recently, such as bitmasks and relative indexing, which are able to encode sparse models, but they are procedural methods that are difficult to implement in parallel
In order to make full use of the resources of deep learning accelerators, a series of novel sparse coding methods have been proposed in recent years, including block compressed sparse column BCSC and nested bit mask NB, which are suitable for parallelization environment, but the compressed model still consumes Massive storage and memory bandwidth
[0006] A major challenge of the above two methods is that it is difficult to achieve high compression ratio and efficient calculation at the same time

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Hierarchical sparse coding method of pruned deep neural network with extremely high compression ratio
  • Hierarchical sparse coding method of pruned deep neural network with extremely high compression ratio
  • Hierarchical sparse coding method of pruned deep neural network with extremely high compression ratio

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0032] Such as figure 1 As shown, in the modified embodiment, a hierarchical sparse coding method for pruned deep neural networks with extremely high compression ratio, including maximizing the compression ratio by greatly reducing the amount of metadata, utilizing a novel and effective The hierarchical sparse coding framework optimizes the coding and inference process design of sparse matrices; proposes a sparse coding method with a very high compression ratio, which uses the LSC method, including two key layers, the block layer and the coding layer. In the block layer, the The sparse matrix is ​​divided into multiple small blocks, and then the zero blocks are deleted. In the encoding layer, a novel SRI method is proposed to further encode these non-zero blocks; moreover, the present invention designs an effective decoding mechanism for the LSC method , to speed up the multiplication of the encoding matrix during the inference stage.

[0033] Specifically include the followi...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention provides a hierarchical sparse coding method of a pruning deep neural network with an extremely high compression ratio. The hierarchical sparse coding method comprises the steps of performing sparse processing on a hyper-parameterized DNN model by applying a pruning technology; providing a layered sparse coding method LSC, and the compression ratio of a trimmed DNN model being improved by designing a coding mechanism of metadata. A multi-process decoding mechanism is designed, so that matrix operation is supported without complete decoding, and the running memory is saved; the hierarchical sparse coding method maximizes the compression ratio by reducing the amount of metadata, and comprises the following steps: dividing a compression process into a block layer and a coding layer; in the block layer, dividing the sparse matrix into a plurality of small blocks, and then deleting zero blocks; in the coding layer, a novel marked relative indexing method SRI being provided tofurther code the non-zero value blocks; the multi-process decoding mechanism accelerating multiplication of a coding matrix in an inference phase; and finally, proving the effectiveness of the proposed LSC method relative to other sparse coding methods through experimental comparison.

Description

technical field [0001] The invention relates to a layered sparse coding method of a pruned deep neural network with a very high compression ratio, belonging to the field of machine learning. Background technique [0002] Deep Neural Networks (DNNs) have evolved into state-of-the-art in many fields, especially in computer vision, natural language processing, and audio processing. However, the massive growth of hidden layers and neurons consumes considerable hardware storage and memory bandwidth, which poses severe challenges for many resource-constrained scenarios in real-world applications. Especially the arrival of the post-Moore era has slowed down the hardware replacement cycle. Specifically, current DNNs have two major bottlenecks: 1. Conflicts with resource-constrained application platforms, such as self-driving tools, mobile phones, mobile robots, and augmented reality AR, which place a heavy burden on the energy consumption and computation of DNN models Quantity is ...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(China)
IPC IPC(8): G06N3/08G06N3/04
CPCG06N3/082G06N3/04
Inventor 李文斌何爱清刘潇霍静姚丽丽高阳
Owner NANJING UNIV
Features
  • R&D
  • Intellectual Property
  • Life Sciences
  • Materials
  • Tech Scout
Why Patsnap Eureka
  • Unparalleled Data Quality
  • Higher Quality Content
  • 60% Fewer Hallucinations
Social media
Patsnap Eureka Blog
Learn More