Unlock instant, AI-driven research and patent intelligence for your innovation.

Sparse activation perception type neural network accelerator based on FPGA

A neural network and accelerator technology, applied in the fields of electronic information and deep learning, can solve the problems of low utilization rate of sparse activation, etc., and achieve the effects of accelerated computing speed, efficient sparse activation, and high data reusability

Pending Publication Date: 2021-02-26
BEIJING UNIV OF TECH
View PDF0 Cites 2 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0004] The purpose of the present invention is to solve the problem that the utilization rate of sparse activation is relatively low in the prior art using the sparse activation neural network accelerator scheme, and proposes a sparse activation-aware neural network accelerator based on FPGA, which realizes a higher on-chip Data multiplexing, while the data allocation method used reduces power consumption and realizes efficient skipping of 0 values ​​​​in sparse activations

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Sparse activation perception type neural network accelerator based on FPGA
  • Sparse activation perception type neural network accelerator based on FPGA
  • Sparse activation perception type neural network accelerator based on FPGA

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0024] Such as figure 1 As shown in the structure, this embodiment relates to a sparse activation-aware neural network accelerator based on FPGA, including a read command generator, a data distributor, a T m operator sub-channels, the size of which is T m × T n The multiply-accumulate array of T n The addition tree group, function module and output buffer composed of three addition trees.

[0025] The read command generator is used to send a read request to the external bus to address the activation and weight data stored in the off-chip memory, and the read request is in accordance with T n The activation and weight of each input channel are performed in units, and the reading order is: feature map from width to height to input channel depth; weights from width to height, and then from input channel depth to output channel depth.

[0026]The data distributor is used for distributing the data read from the off-chip memory to the operation sub-channel in units of input chan...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses a sparse activation perception type neural network accelerator based on an FPGA. The sparse activation perception type neural network accelerator comprises a read command generator, a data distributor, Tm operation sub-channels, a multiply-accumulate array with the size of Tm*Tn, an addition tree group composed of Tn addition trees, a function module and an output cache. The data distributor distributes the data read by an off-chip memory to the operation sub-channels as few as possible. And the operation channel sends the weight and the non-zero activation to the multiply-accumulate array, and the multiply-accumulate array performs operation convolution multiply-accumulate. The invention further provides a sparse activation neural network operation data flow basedon the FPGA. The sparse activation neural network operation data flow reconstructs the position of the non-zero activation value and matches the corresponding weight. The neural network accelerator has high data reusability, data movement is reduced, power consumption is saved, unused operation sub-channels and subsequent modules of the operation sub-channels are closed by a gated clock, power consumption is also saved, sparse activation can be efficiently perceived, most of operations carried out by non-zero activation are enabled to be operations carried out, and the operation speed is increased.

Description

technical field [0001] The invention relates to the fields of electronic information and deep learning, in particular to an FPGA-based highly efficient sparse activation-aware neural network accelerator. Background technique [0002] Convolutional neural network (CNN) has become one of the most popular and effective algorithms in computer vision tasks in recent years. Because of its accuracy rate is significantly higher than traditional algorithms, it has been obtained in the fields of image classification, face recognition, and semantic segmentation. Wide range of applications. As the scale of the CNN network model continues to expand, more computing and storage resources are required. Due to the lack of parallel computing resources of the CPU, the large-scale CNN operation has been unable to do what it wants. On the contrary, the GPU with a large number of stream processors has become the The mainstream platform for CNN training and reasoning, but the defect of high energ...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
IPC IPC(8): G06N3/04G06N3/063
CPCG06N3/063G06N3/045
Inventor 袁海英曾智勇
Owner BEIJING UNIV OF TECH