Configurable approximate multiplier for quantizing convolutional neural network and implementation method of configurable approximate multiplier

A convolutional neural network and multiplier technology, which is applied in the field of configurable approximate multipliers, can solve the problems of poor efficiency, small bit width, and DAS multipliers are not used, and achieves the effect of improving computing efficiency and reducing area overhead.

Active Publication Date: 2020-02-11
ZHEJIANG UNIV
View PDF10 Cites 32 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

However, due to their diagonally shaped critical paths, most multiplier hardware circuits (cf. figure 1 in a) and figure 1 almost 75% in b)) are actually ineffective, thus wasting resources and compromising energy efficiency
Such DAS multipliers ar...

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Configurable approximate multiplier for quantizing convolutional neural network and implementation method of configurable approximate multiplier
  • Configurable approximate multiplier for quantizing convolutional neural network and implementation method of configurable approximate multiplier
  • Configurable approximate multiplier for quantizing convolutional neural network and implementation method of configurable approximate multiplier

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0034] The present invention will be described in further detail below in conjunction with the accompanying drawings and specific embodiments.

[0035] Such as figure 2 As shown, the present invention proposes a configurable approximate multiplier for quantizing convolutional neural networks, including the following modules:

[0036] (1) Sign extension module: will represent the range in -2 n-2 to 2 n-2 The n-bit signed fixed-point number of -1 is expressed as two n / 2-bit signed fixed-point numbers. When the n-bit signed fixed-point number is non-negative, the n / 2-1 bits from the lowest bit to the top are truncated, and in it 0 is added before the highest bit, and the whole is used as the input of the low bit multiplier, and the other n / 2 bits are used as the input of the high bit multiplier.

[0037] When n=8, the split method is:

[0038] 00XX_XXXX=0XXX_XXX→0XXX_0XXX

[0039] When the n-bit signed fixed-point number is negative, if the decimal value is less than -(2 n...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention discloses a configurable approximate multiplier for quantizing a convolutional neural network and an implementation method of the configurable approximate multiplier. The configurable approximate multiplier comprises a symbol extension module, a sub multiplier module and an approximate adder. The symbol extension module splits long-bit-width signed fixed-point number multiplication into two short-bit-width signed fixed-point number multiplication. The sub-multiplier module comprises a plurality of sub-multipliers, each sub-multiplier only receives one signed fixed-point number output by the symbol extension module, and one signed fixed-point number multiplication is completed in combination with the other input; and the approximate adder merges results output by the sub-multiplier modules to obtain a final result of long-bit-width signed fixed-point number multiplication. For two signed fixed-point number multiplication operations with unequal input bit lengths, the speedand the energy efficiency are obviously improved; in a quantitative convolutional neural network with a large number of multiplication operations, the advantages of the method are embodied to the greatest extent.

Description

technical field [0001] The invention relates to engineering technical fields such as low power consumption design, approximate calculation, and convolutional neural network, and in particular to a configurable approximate multiplier for quantizing convolutional neural network and an implementation method thereof. Background technique [0002] Deep learning has achieved great success in the past few years due to its accuracy, robustness, and efficiency in various tasks. Deep learning typically employs convolutional neural network (CNN) architectures that can perform millions to billions of multiply and accumulate (MAC) operations per second. Deep learning is much more computationally intensive than traditional machine learning techniques. Therefore, energy efficiency (i.e., energy consumption per operation) has become key to deep learning implementation and deployment, especially for mobile and embedded devices that want to save energy and meet strict power constraints. [...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
IPC IPC(8): G06F7/53G06N3/063
CPCG06F7/53G06N3/063
Inventor 卓成郭楚亮张力
Owner ZHEJIANG UNIV
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products