Unlock instant, AI-driven research and patent intelligence for your innovation.

Neural network quantification method

A technology of neural network and quantitative method, applied in the direction of neural learning method, biological neural network model, etc., can solve the problems of full-precision network error, full-precision weight value error, neural network precision loss, etc., to improve accuracy and expression Ability, the effect of speeding up training

Pending Publication Date: 2020-11-24
HEFEI UNIV OF TECH
View PDF0 Cites 0 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

Since the weight values ​​of the neural network are mostly floating-point numbers, there will be a large error with the actual full-precision weight values ​​in the network when quantized with a value of ±1. Although the neural network has certain robustness, it will also be different from the full-precision network. A large error is generated, resulting in a loss of the accuracy of the quantized neural network relative to the unquantized neural network

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Examples

Experimental program
Comparison scheme
Effect test

Embodiment 1

[0027] A neural network quantification method specifically comprises the following steps:

[0028] S1, the forward propagation stage of neural network training;

[0029] S2, the backpropagation stage of neural network training;

[0030] S3. S1 and S2 are repeated until the neural network converges, and the quantization of the deep neural network is completed.

[0031] In S1, before neural network training, an array of quantized weight values ​​with low bits as indexes is initialized as a full-precision quantized value storage model.

[0032] In S1, the specific process is: first quantize the weight value of the current layer network with a quantization function, then calculate the output of the current layer network, store the weight value matrix with a low-bit index value to refer to the corresponding full-precision quantization value, and calculate the first layer of the neural network Loop through to the last layer.

[0033] In S2, the following steps are specifically in...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses a neural network quantization method, belongs to the technical field of deep learning training algorithms, and particularly comprises a forward propagation stage of neural network training and a back propagation stage of neural network training, and the above processes are repeated until a neural network converges to complete deep neural network quantization. Compared witha non-quantized network, the precision loss of the quantized deep neural network can be ignored, the storage capacity and the operation complexity of the model can be reduced, and the quantized deep neural network can be conveniently transplanted to embedded hardware equipment to play the calculation advantage of the quantized neural network on hardware. According to the neural network quantized by the quantization method, a high-precision quantization value can be expressed by using low-bit coding; according to the method, the model memory space and the operation complexity can be effectivelyreduced, the high precision of the network cannot be reduced, the method is more suitable for a hardware shift calculation mode, and the method can be deployed on corresponding hardware to fully exert the calculation advantages of the quantitative neural network.

Description

technical field [0001] The present invention relates to the technical field of deep learning training algorithms, and more specifically, relates to a neural network quantization method. Background technique [0002] Artificial neural network, also referred to as neural network or connection model for short, is an algorithmic mathematical model that imitates the behavior characteristics of animal neural networks and performs distributed parallel information processing. This kind of network depends on the complexity of the system, and achieves the purpose of processing information by adjusting the interconnection relationship between a large number of internal nodes. Artificial neural network is a mathematical model that uses a structure similar to that of the brain's synaptic connections for information processing. In engineering and academia, it is often referred to directly as "neural network" or neural network. [0003] The existing method of quantizing the neural networ...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(China)
IPC IPC(8): G06N3/08
CPCG06N3/084
Inventor 樊春晓胡洲宋光明王振兴
Owner HEFEI UNIV OF TECH