Sparse neural network processor based on systolic array

A neural network and processor technology, applied in the field of computer architecture for sparse neural network computing, can solve problems such as irregularity of sparse patterns, and difficulty in effectively utilizing the sparsity of neural networks, reducing memory capacity, reducing space overhead, The effect of removing data redundancy

Active Publication Date: 2020-01-17
BEIHANG UNIV
View PDF6 Cites 12 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

However, it is worth noting that due to the irregularity of the sparse patterns of neural network weights, gradients, and features, this sparsity of neural networks is often difficult to be effectively utilized, especially on highly regular computing platforms such as systolic arrays.

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Sparse neural network processor based on systolic array
  • Sparse neural network processor based on systolic array
  • Sparse neural network processor based on systolic array

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0036] In order to make the purpose, technical solution and advantages of the present invention clearer, the present invention will be further described in detail in combination with specific embodiments and accompanying drawings.

[0037] figure 1 is a schematic diagram of the architecture of a sparse neural network processor according to an embodiment of the present invention. Including storage unit, control unit, bus array, sparse matrix operation array and computing unit. The storage unit is used to store neural network weights, gradients, features, and instruction sequences for data flow scheduling. The control unit is connected to the storage unit, and according to the scheduling of the instruction sequence, obtains the required data from the storage, and reorganizes the data into the form of matrix operation, and then sends it to the sparse matrix operation array after being bridged by the bus array to complete the corresponding calculations. The calculation unit rec...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention provides a sparse neural network processor based on a systolic array. The sparse neural network processor comprises a storage unit, a control unit, a sparse matrix operation array, a calculation unit and a confluence array. The storage unit is used for storing weights, gradients, features and instruction sequences used for scheduling data streams. The control unit takes out data required by the training and reasoning process from the storage unit according to the control of the instruction sequence, converts the data into a sparse matrix operation format and sends the data into the sparse matrix operation array. The sparse matrix operation array comprises a plurality of processing units connected in a systolic array mode and is used for completing sparse matrix operation. Thecalculation unit is used for completing element-by-element operation such as a nonlinear activation function. The confluence array delivers the same data segment to different rows of the systolic array through internal data transfer to reduce storage overhead. The processor makes full use of the sparsity of the weight and the characteristics, achieves the improvement of the speed and power consumption ratio in the neural network training and reasoning process, and has the advantages of high concurrency, low bandwidth requirements and the like.

Description

technical field [0001] The invention relates to neural network technology and computer architecture, in particular to computer architecture for sparse neural network calculation. Background technique [0002] In recent years, deep learning has gradually achieved more and more remarkable results in fields such as image recognition and speech processing. However, as the depth of the network continues to increase, the computing power and memory access bandwidth required in the training and inference process of the deep neural network are gradually difficult to be met by traditional computing platforms. Therefore, various domain-specific architectures (domain specified architectures) applied to neural networks have been proposed by the industry and academia to meet this requirement. Among them, the systolic array architecture has attracted great attention from the industry and academia due to its characteristics of high concurrency and low bandwidth requirements. [0003] On t...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
Patent Type & Authority Applications(China)
IPC IPC(8): G06N3/063G06F15/80G06N3/04
CPCG06N3/063G06F15/8046G06N3/048
Inventor 杨建磊赵巍胜付文智
Owner BEIHANG UNIV
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products