Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Neural network accelerator for bit width partitioning and implementation method of neural network accelerator

A technology of neural network and implementation method, which is applied in the field of computer vision, can solve the problem of sacrificing hardware area indicators, and achieve the effects of increasing DSP utilization, increasing resource utilization efficiency, and improving computing throughput

Active Publication Date: 2017-12-08
TSINGHUA UNIV
View PDF2 Cites 41 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

However, this design method converts the multiplication and accumulation operation in the convolutional neural network into an AND or logic operation, which improves the computing performance of the accelerator, but increases the hardware area by 35%, but largely sacrifices the CNN hardware. Area index

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Neural network accelerator for bit width partitioning and implementation method of neural network accelerator
  • Neural network accelerator for bit width partitioning and implementation method of neural network accelerator
  • Neural network accelerator for bit width partitioning and implementation method of neural network accelerator

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0035] The following will clearly and completely describe the technical solutions in the embodiments of the present invention with reference to the accompanying drawings in the embodiments of the present invention. Obviously, the described embodiments are only some, not all, embodiments of the present invention. Based on the embodiments of the present invention, all other embodiments obtained by persons of ordinary skill in the art without making creative efforts belong to the protection scope of the present invention.

[0036] In the prior art, mainstream CNN accelerators use a single bit-width computing processing unit (CP) to iteratively calculate each layer of CNN. However, due to the different data bit width requirements of each layer of CNN, such a design will lead to a lot of problems. Big waste of resources. Based on this, the embodiment of the present invention provides a neural network accelerator for bit width partitioning and its implementation method. According to...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The present invention provides a neural network accelerator for bit width partitioning and an implementation method of the neural network accelerator. The neural network accelerator includes a plurality of computing and processing units with different bit widths, input buffers, weight buffers, output buffers, data shifters and an off-chip memory; each of the computing and processing units obtains data from the corresponding input buffering area and weight buffer, and performs parallel processing on data of a neural network layer having a bit width consistent with the bit width of the corresponding computing and processing unit; the data shifters are used for converting the bit width of data outputted by the current computing and processing unit into a bit width consistent with the bit width of a next computing and processing unit corresponding to the current computing and processing unit; and the off-chip memory is used for storing data which have not been processed and have been processed by the computing and processing units. With the neural network accelerator for bit width partitioning and the implementation method of the neural network accelerator of the invention adopted, multiply-accumulate operation can be performed on a plurality of short-bit width data, so that the utilization rate of a DSP can be increased; and the computing and processing units (CP) with different bit widths are adopted to perform parallel computation of each layer of a neural network, and therefore, the computing throughput of the accelerator can be improved.

Description

technical field [0001] The invention relates to computer vision technology, in particular to a neural network accelerator for bit width partitioning and its implementation method. Background technique [0002] In recent years, convolutional neural networks have made great progress in many computer vision applications, such as image recognition, video classification, gesture recognition, etc. Recently, with the introduction of more and more convolutional neural network (CNN) algorithms, the recognition accuracy of CNN has been greatly improved, and the recognition accuracy reached 96.4% at the 2015 ILSVRC conference. However, in order to achieve such a high recognition accuracy, CNN needs to introduce very complex algorithms, which will increase the overhead of hardware computing resources and greatly limit the performance improvement of CNN at the hardware level. Therefore, many acceleration methods for CNN have been proposed to increase the computing performance of CNN, su...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(China)
IPC IPC(8): G06N3/063
CPCG06N3/063
Inventor 尹首一郭建辛欧阳鹏唐士斌涂锋斌刘雷波魏少军
Owner TSINGHUA UNIV
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products