Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

A sparse neural network accelerator and its implementation method

A technology of neural network and implementation method, which is applied in the field of sparse neural network accelerator and its implementation, can solve the problem of large memory power consumption, and achieve the effect of reducing the total size

Active Publication Date: 2022-02-01
TSINGHUA UNIV
View PDF11 Cites 0 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0006] In order to overcome the above-mentioned problem of high power consumption of the existing output memory or at least partially solve the above-mentioned problem, an embodiment of the present invention provides a sparse neural network accelerator and its implementation method

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • A sparse neural network accelerator and its implementation method
  • A sparse neural network accelerator and its implementation method
  • A sparse neural network accelerator and its implementation method

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0041] In order to more clearly illustrate the technical solutions in the embodiments of the present invention or the prior art, the following will briefly introduce the drawings that need to be used in the description of the embodiments or the prior art. Obviously, the accompanying drawings in the following description are the For some embodiments of the invention, those skilled in the art can also obtain other drawings based on these drawings without creative effort.

[0042] In one embodiment of the present invention, a sparse neural network accelerator is provided, figure 1 The sparse neural network accelerator provided for the embodiment of the present invention includes a PE array and an output memory, and the PE array is divided into multiple PE groups, wherein each PE group and the corresponding output memory form an association groups, the number of PE units in each associated group is equal to the number of output memories. For each PE unit in any association group,...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The present invention provides a sparse neural network accelerator and its implementation method. The accelerator mainly includes a PE array, an output memory, and a scheduler module. The PE array is divided into multiple PE groups, and each PE group and the corresponding output memory form an associated group. The number of PEs in the association group is equal to the number of output memories; each PE unit in the association group can access any output memory in the association group; PE units in any association group are calculated according to the input activation value and weight value Multiple output results are obtained and written into corresponding multiple output memories according to preset rules; the scheduler module schedules the sequence of output activation values ​​to reduce the probability of hash collisions. The invention divides the original PE array into multiple PE groups, and forms an associated group architecture with the corresponding output memory, which greatly reduces the area of ​​the output memory and reduces power consumption; the scheduler module reduces the probability of hash conflicts and greatly improves the Computational performance of the entire system.

Description

technical field [0001] The invention belongs to the technical field of neural network accelerators, in particular to a sparse neural network accelerator and an implementation method thereof. Background technique [0002] Convolutional Neural Networks (CNNs) have shown promising performance in many domains. However, CNN networks have considerable computational and storage complexity, so application-specific integrated circuit (ASIC) accelerators need to be designed to speed up the computation of CNNs. At the same time, existing high-efficiency CNNs have high sparsity (the proportion of 0 in network parameters and intermediate results is high), so existing accelerators adopt different hardware designs to accelerate the calculation of sparse neural networks. [0003] Existing ASIC designs have taken different approaches to handle sparse CNNs. The first method is the power gating method, which improves energy efficiency by turning off the corresponding computing unit (PE) when...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Patents(China)
IPC IPC(8): G06N3/04G06N3/063G06N3/08
CPCG06N3/063G06N3/082G06N3/045
Inventor 刘勇攀王靖宇袁哲杨华中
Owner TSINGHUA UNIV
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products