A CNN-based low-precision training and 8-bit integer quantitative reasoning method

A reasoning method and low-precision model technology, applied in the field of convolutional neural networks, can solve the problems of large precision loss and insufficient calculation efficiency, and achieve the effects of controlling precision loss, saving calculation time, and reducing precision loss

Inactive Publication Date: 2019-06-18
成都康乔电子有限责任公司 +1
View PDF0 Cites 60 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0005] The problem to be solved by the present invention is to overcome the defects in the above-mentioned prior art, and propose a CNN-based low-precision training and 8-bit integer quantization reasoning method to solve the large loss of precision existing in the existing quantization method , Calculation efficiency is not high enough

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • A CNN-based low-precision training and 8-bit integer quantitative reasoning method
  • A CNN-based low-precision training and 8-bit integer quantitative reasoning method
  • A CNN-based low-precision training and 8-bit integer quantitative reasoning method

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0027] It should be noted that, in the case of no conflict, the embodiments of the present invention and the features in the embodiments can be combined with each other.

[0028] Specific embodiments of the invention will be described in detail below.

[0029] The technical solution of the present invention is divided into two stages: the first stage is to use 16-bit floating-point low-precision fixed-point algorithm for model training to obtain a model for target detection, that is, weights. The second stage is to use the 8-bit integer quantization scheme to quantize the weights, quantize the activation value into 8-bit integer data, and realize the quantitative reasoning of 8-bit integers.

[0030] Specific steps are as follows:

[0031] A. Use the 16-bit floating-point low-precision fixed-point algorithm of to train the model, that is, use 2 bits to represent the integer part, use 14 bits to represent the fractional part, and use rounding to convert 32-bit floating-point ...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention provides a CNN-based low-precision training and 8-bit integer quantization reasoning method. The method mainly comprises the steps of carryin gout low-precision model training; Performing model training by using a 16-bit floating point type low-precision fixed point algorithm to obtain a model for target detection; Quantifying the weight; Proposing an 8-bit integer quantization scheme, and quantizing the weight parameters of the convolutional neural network from 16-bit floating point type to 8-bit integer according to layers; carrying out 8-bit integer quantitative reasoning; quantizing the activation value into 8-bit integer data, i.e., each layer of the CNN accepts an int8 type quantization input and generates an int8 quantization output. According to the invention, a 16-bit floating point type low-precision fixed point algorithm is used to train a model to obtain a weight; Compared with a 32-bit floating point type algorithm, the method has the advantages that the 8-bit integer quantization reasoning is directly carried out on the weight obtained by training the model, the reasoning process of the convolutional layer is optimized, and the precision loss caused by the low-bit fixed point quantization reasoning is effectively reduced.

Description

technical field [0001] The invention belongs to the technical field of convolutional neural network, and in particular relates to a CNN-based low-precision training and 8-bit integer quantization reasoning method. Background technique [0002] Convolutional Neural Networks (CNNs) have achieved excellent results in the fields of image classification, object detection, face recognition, etc. To realize real-time forward reasoning of CNNs on the embedded platform, it is necessary to compress the model size of the neural network and improve the computational efficiency of the model under the condition of controlling the loss of accuracy. [0003] The current commonly used method is to quantize the weight and (or) activation value of CNN, and convert the data from 32-bit floating point type to lower integer type. However, the current quantization methods still have shortcomings in the trade-off between accuracy and computational efficiency. Many quantization methods perform netw...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
Patent Type & Authority Applications(China)
IPC IPC(8): G06K9/62G06K9/00G06N3/08G06N3/04G06N5/04
Inventor 严敏佳王永松刘丹
Owner 成都康乔电子有限责任公司
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products