Unlock instant, AI-driven research and patent intelligence for your innovation.

Compressed encoding method of sparse neural network

A neural network and compression coding technology, applied in the field of compression coding of sparse neural network, can solve the problems of complex coding and achieve the effect of saving bandwidth

Active Publication Date: 2019-06-07
HANGZHOU NATCHIP SCI & TECH
View PDF4 Cites 10 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

Most of the existing technologies directly use Huffman coding. This method is more complicated to encode, and at the same time, it is necessary to save a code table before storing or transmitting the coded data for use when the decoder reconstructs the information.

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Compressed encoding method of sparse neural network
  • Compressed encoding method of sparse neural network
  • Compressed encoding method of sparse neural network

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0035] The present invention will be further described below in conjunction with the accompanying drawings and embodiments. It should be pointed out that this implementation method is only used to explain the present invention, and does not limit the implementation scenarios of the present invention.

[0036] Such as figure 1 , a compression coding method for sparse neural networks, first quantize and preprocess the weights and activation data in the neural network, and select the compression coding method according to the preprocessed weights and data sparsity: sparsity S≥ε 2 When , use zero-run and k-order exponential Columbus combination coding; sparsity ε 1 2 When , use k-order GX coding; sparsity S≤ε 1 When , the k-order exponential Columbus code is used; ε 1 and ε 2 To set the threshold, 0≤ε 1 2 ≤1.

[0037] This embodiment adopts the open source pre-training model ResNet V2_50 of tensorflow on github, and the download address is https: / / github.com / tensorflow / model...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention relates to a compressed encoding method of a sparse neural network. The method comprises the following steps: firstly, carrying out quantization and preprocessing, and selecting a compression coding method according to a sparsity degree: when the sparsity S is greater than or equal to 2, zero-run and k-order exponential Golomb combined coding is adopted; when the sparsity Epsilon 1<S<Epsilon 2, k-order GX coding is adopted; when the sparsity S is smaller than or equal to Epsilon1, k-order exponential Golomb coding is adopted; wherein the Epsilon1 and the Epsilon2 are set threshold values. The k-order GX coding method comprises the following steps of: if the preprocessing completion data is 0, directly coding into 1 in a binary form, if the preprocessing completion data is greater than 0, representing in the binary form, insufficient k bits and high-order zero padding, and removing the low k bits of the binary bit sequence to obtain a sequence which is converted into a decimal number y; if k bits exist after high-order zero padding, y is 0; calculating a least significant digit LSB in a y + 1 binary form, outputting LSB 0, then outputting a y + 1 bit sequence, and putting the k-bit-removed binary sequence at the output least bit to obtain a coded code word;. According to the method, the distribution characteristics of the sparse neural network weight matrix are utilized, so that the method has higher compression ratio and lower implementation complexity.

Description

technical field [0001] The invention belongs to the technical field of computers, in particular to the field of neural networks, and relates to a compression coding method of a sparse neural network. Background technique [0002] With the advent of the era of artificial intelligence, intelligent tasks such as image recognition, speech recognition, and natural language processing are ubiquitous in life. As one of the most effective algorithms for realizing such intelligent tasks, the neural network has received extensive attention and application in academia and industry. However, a large neural network has a large number of layers and nodes, resulting in a large number of weight parameters, time-consuming network training process, and a trained model takes up a large storage space. The computation-intensive and storage-intensive characteristics of neural networks make it difficult to deploy to mobile and embedded systems with limited resources. Therefore, neural network co...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
IPC IPC(8): G06T9/00H03M7/30G10L19/00
Inventor 莫冬春钟宇清黄磊杨常星宋蕴胡俊陈伟钟天浪
Owner HANGZHOU NATCHIP SCI & TECH