Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

2-exponential power deep neural network quantification method based on knowledge distillation training

A technology of deep neural network and quantization method, which is applied in the field of 2-exponential deep neural network quantization based on knowledge distillation training, can solve problems such as error and neural network accuracy loss of neural network, so as to reduce error, improve accuracy, and improve calculation. The effect of efficiency

Pending Publication Date: 2020-11-24
HEFEI UNIV OF TECH
View PDF0 Cites 4 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0005] Aiming at the deficiencies of the prior art, the present invention provides a 2-exponential power deep neural network quantization method based on knowledge distillation training, which solves the problem that the gradient of the quantized weight value adopts full precision when the network performs backpropagation in the existing method of training the quantized neural network. The gradient accumulated by the weight value is approximated, and the quantized weight value of the neural network will have an error with the actual full-precision weight value in the network. Although the neural network has certain robustness, this will also lead to large errors, resulting in its The problem that the accuracy of the quantized neural network is lost relative to the unquantized neural network

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • 2-exponential power deep neural network quantification method based on knowledge distillation training

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0026] The following will clearly and completely describe the technical solutions in the embodiments of the present invention with reference to the accompanying drawings in the embodiments of the present invention. Obviously, the described embodiments are only some, not all, embodiments of the present invention. Based on the embodiments of the present invention, all other embodiments obtained by persons of ordinary skill in the art without making creative efforts belong to the protection scope of the present invention.

[0027] Such as figure 1 As shown, the present invention provides a technical solution: a 2-exponential power deep neural network quantization method based on knowledge distillation training, including a teacher model and a 2-exponential power quantized student model, and the teacher network model has more parameters and a higher precision network The student model generally chooses a network model with fewer parameters and less precision than the teacher model...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention relates to the technical field of neural networks. The invention further discloses a 2-exponential power deep neural network quantification method based on knowledge distillation training. The method comprises a teacher model and a student model with exponential power quantification of 2, and is characterized in that the teacher network model selects a network model with more parameters and higher precision, and the student model generally selects a network model with fewer parameters and lower precision than the teacher model. According to the invention, an exponential power quantification deep neural network method in which a neural network weight value is quantified into 2 is adopted; an error with a full-precision weight value can be reduced; the precision of the trainednetwork and the precision loss of the unquantified network are effectively reduced; moreover, the exponential power weight multiplication operation of 2 can be completed by displacement, the method has obvious calculation advantages in hardware equipment deployment, the calculation efficiency of neural network hardware can be improved, and the neural network model trained based on the knowledge distillation algorithm can effectively improve the accuracy of the quantitative network.

Description

technical field [0001] The invention relates to the technical field of neural networks, in particular to a method for quantifying a power-of-two deep neural network based on knowledge distillation training. Background technique [0002] Artificial neural network, also referred to as neural network or connection model for short, is an algorithmic mathematical model that imitates the behavior characteristics of animal neural networks and performs distributed parallel information processing. This kind of network depends on the complexity of the system, and achieves the purpose of processing information by adjusting the interconnection relationship between a large number of internal nodes. [0003] The existing methods for training quantized neural networks are mainly to quantify the weight value of each layer to a low-precision value (such as +1, -1, 0, etc. integers) in the forward propagation stage of the neural network, and then calculate its The output of the layer network...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(China)
IPC IPC(8): G06K9/62G06N3/08
CPCG06N3/08G06F18/24
Inventor 樊春晓胡洲宋光明王振兴
Owner HEFEI UNIV OF TECH
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products