Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Convolutional neural network weight compression method and device based on arm architecture fpga hardware system

A convolutional neural network and ARM architecture technology, applied in the field of neural network speed-up, can solve problems such as poor implementation effect, no hardware characteristics taken into account, and increased computing system burden, so as to improve computing efficiency, optimize storage, and reduce the number of multiplications Effect

Active Publication Date: 2020-12-08
SHANGHAI ANLOGIC INFOTECH CO LTD
View PDF9 Cites 0 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

However, Huffman coding is not for huge sparse matrices. For the convolution calculation process of convolutional neural networks with huge sparse matrices, Huffman coding needs to count the number of weights before calculating the Huffman tree to get Long encoding, which undoubtedly increases the burden on the computing system
At present, the process of Huffman encoding and decoding does not take hardware characteristics into account, and the implementation effect on FPGA is not good.

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Convolutional neural network weight compression method and device based on arm architecture fpga hardware system
  • Convolutional neural network weight compression method and device based on arm architecture fpga hardware system
  • Convolutional neural network weight compression method and device based on arm architecture fpga hardware system

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0040] In the following description, many technical details are proposed in order to enable readers to better understand the application. However, those skilled in the art can understand that the technical solutions claimed in this application can be realized even without these technical details and various changes and modifications based on the following implementation modes.

[0041] Explanation of some concepts:

[0042] Sparse Matrix: The number of non-zero elements in the matrix is ​​much smaller than the total number of matrix elements, and the distribution of non-zero elements is irregular. It is generally considered that the total number of non-zero elements in the matrix is ​​less than the total value of all elements in the matrix When it is equal to 0.05, the matrix is ​​called a sparse matrix.

[0043] Huffman Encoding (Huffman Encoding): It is an entropy coding (weight coding) greedy algorithm for lossless data compression.

[0044] In order to make the purpose, ...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The present application relates to speed-up of neural networks, and discloses a convolutional neural network weight compression method and a device thereof. The method includes obtaining all the convolution kernels of the convolutional neural network; each weight matrix for any convolution kernel is stacked and arranged according to the depth from small to large, and all the arranged multi-layer weight matrices except the first Any non-zero element of : Calculate the depth offset of the current non-zero element relative to the previous non-zero element and calculate the height offset of the current non-zero element relative to the first row and first column element of the current layer weight matrix and width offset, and compress the value of the current non-zero element, the depth offset, the height offset, and the width offset according to the preset compression rules to obtain the corresponding compression result. The embodiment of the present application enables the compressed result to accelerate the subsequent neural network operation on the FPGA.

Description

technical field [0001] The present application relates to the field of neural network speed-up, in particular to convolutional neural network weight compression technology. Background technique [0002] For non-uniformly distributed information, Huffman coding is the theoretically optimal lossless compression scheme. However, Huffman coding is not aimed at huge sparse matrices. For the convolution calculation process of convolutional neural networks with huge sparse matrices, Huffman coding needs to count the number of weights before calculating the Huffman tree to get Long encoding, which undoubtedly increases the burden on the computing system. At present, the process of Huffman encoding and decoding does not take hardware characteristics into consideration, and the implementation effect on FPGA is not good. Contents of the invention [0003] The purpose of this application is to provide a convolutional neural network weight compression method and device thereof, so th...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Patents(China)
IPC IPC(8): G06N3/04G06N3/08
CPCG06N3/082G06N3/045
Inventor 边立剑叶梦琦
Owner SHANGHAI ANLOGIC INFOTECH CO LTD
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products