Check patentability & draft patents in minutes with Patsnap Eureka AI!

Weight and activation value quantification method for long-term and short-term memory network

A technology of long-term short-term memory and quantization method, which is applied in the quantization field of weights and activation values, which can solve problems such as large amount of calculation and limitation of neural network application scenarios, and achieve the effect of reducing data bit width and reducing errors

Pending Publication Date: 2020-12-22
XI AN JIAOTONG UNIV
View PDF0 Cites 2 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

However, it is difficult for desktop processors and mobile processors to bear such a large amount of calculation, which greatly limits the application scenarios of neural networks, including various terminal applications such as virtual reality technology and augmented reality technology

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Weight and activation value quantification method for long-term and short-term memory network
  • Weight and activation value quantification method for long-term and short-term memory network
  • Weight and activation value quantification method for long-term and short-term memory network

Examples

Experimental program
Comparison scheme
Effect test

Embodiment

[0056] The present invention can be realized in the software and hardware system of the long short-term memory network.

[0057] Taking the input parameters of the LSTM network as an example, quantify the weight value, a total of three steps:

[0058] The first step is to search for a suitable target quantization range num. During the experiment, a total of 2 N values, where N=1~7, run the kl_comp_weight() function to collect the KL divergence under the optimal quantization scheme for each num, the result is as follows Figure 6 shown. When num increases, the KL divergence decreases, that is, when the bit width of the fixed-point number is larger, the distribution of the fixed-point number and the floating-point number is closer, and the error is smaller. When num is greater than 8, the KL divergence decreases rapidly as num increases. In order to reduce the cost of hardware design, the quantization of input parameters is selected as INT3.

[0059] In the second step, in ...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses a weight and activation value quantification method for a long-term and short-term memory network. The method comprises the following steps: 1) collecting a weight and activation value number set of the long-term and short-term memory network; 2) determining a corresponding target quantization range, starting threshold traversal circulation, and scaling or setting a weightvalue and an activation value as saturation values according to the calculated scaling factors and saturation values under different threshold conditions; and 3) after traversing is completed, respectively calculating the KL divergence of the initial number set of the weight value and the activation value and the KL divergence of the mapped number set, and finally respectively outputting positiveand negative number direction truncation thresholds and the minimum KL divergence value. According to the method, the long-term and short-term memory network obtained after high-precision floating-point number training is completed is converted into the fixed-point number network, a quantitative structure for the weight value and the activation value of the long-term and short-term memory networkis innovatively designed, the hardware overhead is reduced while the algorithm hardware implementation precision is guaranteed, and the operation speed is increased.

Description

technical field [0001] The invention belongs to the field of cyclic neural network quantization, and in particular relates to a quantization method for weight and activation values ​​of long-term and short-term memory networks. Background technique [0002] With the continuous improvement of the computing power of graphics processors and general-purpose central processing units, the requirements for computational complexity of artificial neural networks have been eased. After 2012, artificial intelligence algorithms based on neural networks have been continuously developed and widely used in many fields such as pattern recognition, speech processing, and image processing. However, the development of hardware performance cannot satisfy the evolution of algorithms. The SSD network proposed in 2016 introduces up to 50 billion calculations and needs to be run in large workstations. However, desktop processors and mobile processors cannot bear such a large amount of calculation...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
IPC IPC(8): G06N3/04G06N3/08
CPCG06N3/08G06N3/048G06N3/045G06N3/044
Inventor 杨晨丰贵鹏王逸洲耿莉
Owner XI AN JIAOTONG UNIV
Features
  • R&D
  • Intellectual Property
  • Life Sciences
  • Materials
  • Tech Scout
Why Patsnap Eureka
  • Unparalleled Data Quality
  • Higher Quality Content
  • 60% Fewer Hallucinations
Social media
Patsnap Eureka Blog
Learn More