Weight value fixed-point quantification method for lightweight convolutional neural network
A technology of convolutional neural network and quantization method, which is applied in the field of fixed-point quantization of weight values, can solve the problems of accuracy loss, low parameter redundancy, and convolutional neural network accuracy loss, and achieves low accuracy loss and low hardware. The effect of resource consumption, fast retraining speed
- Summary
- Abstract
- Description
- Claims
- Application Information
AI Technical Summary
Problems solved by technology
Method used
Image
Examples
Embodiment
[0053] In this research field, Google proposed to introduce a pseudo-quantization method in the training process for the training process of the network, to realize the algorithm of directly training the quantized network, and the final quantization factor used fixed-point numbers to complete the integerization of the entire network. The tensorRT tool launched by NVIDIA, through the verification of 500 pictures, directly converts a pre-trained convolutional neural network model into an integer with only 8 bits of weight, thereby converting floating-point multiplication into 8-bit integer operations, without the need for any more. training, but the method's quantization and inverse quantization relies on floating-point multiplication. Intel China Research Institute proposed an incremental quantization and training process for neural networks, gradually transforming the model into a network with a weight of only a power of 2 or 0, and maintaining a good accuracy rate.
[0054]Th...
PUM
Login to View More Abstract
Description
Claims
Application Information
Login to View More 


