Hardware architecture of binary weight convolution neural network accelerator and calculation process thereof

A binary weight convolution and neural network technology, applied to the hardware architecture and calculation process of the binary weight convolutional neural network dedicated accelerator, to achieve the effect of reducing access, reducing access, and reducing power consumption requirements

Active Publication Date: 2017-06-20
南京风兴科技有限公司
View PDF4 Cites 83 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0005] The present invention aims to solve the technical problem of applying convolutional ne...

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Hardware architecture of binary weight convolution neural network accelerator and calculation process thereof
  • Hardware architecture of binary weight convolution neural network accelerator and calculation process thereof
  • Hardware architecture of binary weight convolution neural network accelerator and calculation process thereof

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0039] The embodiments of the present invention are detailed below, and examples of the embodiments are shown in the accompanying drawings. First introduce the necessary hardware overall architecture, and then introduce the optimized calculation process based on this hardware architecture. The following implementations described with reference to the accompanying drawings are exemplary, and are intended to explain the present invention, but should not be construed as limiting the present invention.

[0040] In the description of the present invention, it should be understood that the orientation or positional relationship indicated by the terms "upper", "lower", "left", "right", "vertical", "horizontal", etc. are based on the drawings shown The orientation or positional relationship of is only a simplified description for the convenience of describing the present invention, rather than indicating or implying that the device or element referred to must have a specific orientation,...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention discloses the hardware architecture of a binary weight convolution neural network accelerator and a calculation process thereof. The hardware architecture comprises three double-ended on-chip static random access memories which are used for buffering the binary weight of input neurons and a convolution layer, four convolution processing units capable of controlling calculation parts to complete major convolution calculation operation according to the calculation process, a feature map accumulation unit and a convolutional accumulation array. The feature map accumulation unit and the convolutional accumulation array are used for further processing the operation result of the convolution processing units to acquire a final correct output neuron value. The entire design exchanges data with an off-chip memory via a dynamic random access memory interface. In addition to the hardware architecture, the invention further provides the detailed calculation process which optimizes the hardware architecture and uses four lines of input feature map as a complete calculation unit. According to the invention, input data are reused to the greatest extent; the access of the off-chip memory is eliminated as much as possible; the power consumption of the deep binary convolution neural network calculation can be effectively reduced; a deep network is supported; and the scheme provided by the invention is a reasonable scheme which can be applied to an embedded system of visual application.

Description

Technical field [0001] The present invention is designed in the field of computer and electronic information technology, and particularly relates to a hardware architecture of a special accelerator for binary weighted convolutional neural networks and its calculation process. Background technique [0002] The deep convolutional neural network model has achieved great breakthroughs and success in many fields such as image classification, motion detection, speech recognition and other big data analysis tasks. On the one hand, as the effect of the convolutional neural network becomes better and better, its topological structure is constantly deepening, and the number of parameters has reached the 6th power of 10 and above, which brings extreme computational complexity. Big upgrade, explosive growth in computing power required. On the other hand, embedded systems can only provide limited resources, and their power consumption is also limited within a certain range. Although the exis...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
IPC IPC(8): G06N3/063
CPCG06N3/063
Inventor 王中风王逸致林军
Owner 南京风兴科技有限公司
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products