Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Sparse neural network-oriented system on chip

A neural network and system-on-a-chip technology, applied in biological neural network models, complex mathematical operations, instruments, etc., can solve problems such as difficulty in data reuse of accelerators, low CPU utilization, and general acceleration effects, and achieve improved data utilization, The effect of improving efficiency and reducing the number of memory accesses

Active Publication Date: 2021-04-09
UNIV OF ELECTRONICS SCI & TECH OF CHINA
View PDF8 Cites 0 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0005] b. Uncertainty in the position of non-zero elements in the sparse network leads to difficulties in data reuse in the accelerator;
[0006] c. It is necessary to design a special compiler for the accelerator, and the versatility is poor
[0007] In the system on chip, the utilization rate of the CPU is too low, and the overall acceleration effect of the system is mediocre.

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Sparse neural network-oriented system on chip
  • Sparse neural network-oriented system on chip
  • Sparse neural network-oriented system on chip

Examples

Experimental program
Comparison scheme
Effect test

Embodiment 1

[0049] A system-on-a-chip for sparse neural networks, in this embodiment, includes the main processor type open source processor Rocket, the neural network accelerator of coprocessor type, system bus TileLink, debug interface UART, SPIFlash controller and chip external SPI Flash memory. Among them, the processor is the main device of the system; the neural network accelerator is the coprocessor of the main processor, and the coprocessor is connected to the main processor through the RoCC interface; the UART and SPI Flash controller are the slave devices of the system, and the slave device is connected through TileLink The bus is connected to the main processor; SPI Flash is integrated into the system as an off-chip storage through the SPI interface in the SPI Flash controller. The processor is mainly responsible for decomposing the matrix calculation in the neural network algorithm into vector calculation, and then filtering out the non-zero data participating in the calculati...

Embodiment 2

[0051] Such as figure 1 As shown, this embodiment is a heterogeneous system-on-chip created based on a RISC-V open source processor. The system integrates Rocket CPU, coprocessor accelerator, SPI Flash controller, UART and off-chip SPI Flash. RocketCPU serves as the main processor of the system, and the accelerator that executes the neural network algorithm serves as its coprocessor. The main processor and the coprocessor are tightly coupled through the RoCC interface, and the two share the L1 DCache. SPI Flash is an off-chip program memory, externally connected to the SPI Flash controller, which stores the binary file of the program inside. After the system is powered on and reset, the CPU reads a block of instructions from the base address of the SPI Flash to the L1 ICache. . Then, the CPU executes the program to filter and reorganize the non-zero neurons and non-zero weights involved in the calculation in the sparse neural network, and convert the sparse vector in the spa...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses a sparse neural network-oriented system on chip, which comprises a main processor, a co-processor, system slave equipment and an off-chip memory, the main processor is in communication connection with the main processor; the system slave equipment is in communication connection with the main processor; the slave system interface of the off-chip memory is in communication connection with the slave system; the main processor decomposes matrix calculation in a neural network algorithm into vector calculation, executes a program to screen and recombine sparse vectors participating in calculation in a sparse neural network, converts the sparse vectors in the sparse neural network into dense vectors, and sends an acceleration instruction to the co-processor. Accelerated computation of dense vectors is performed by the co-processor. According to the invention, the problems that the accelerator parallelism and the memory access efficiency are limited and the utilization efficiency of a processor in the system is low are solved. For the direct index storage of the sparse neural network, the problem of difficult data reuse in the accelerator caused by nonzero element position uncertainty in the sparse neural network is solved.

Description

technical field [0001] The invention relates to the communication field, in particular to a system on chip oriented to a sparse neural network. Background technique [0002] The sparseness of the neural network greatly reduces the amount of calculation and data storage of the algorithm, which helps to deploy large-scale neural network algorithms in embedded devices with limited storage, computing resources and energy consumption. However, the data irregularity caused by sparseness makes the execution efficiency of sparse networks on general platforms very low. In order to make hardware perform sparse neural network algorithms better, researchers have designed a variety of sparse neural network accelerators. In addition, the research on SoCs for sparse neural networks has gradually begun. [0003] There are several problems with existing sparse neural network accelerators and SoCs: [0004] a. The calculation parallelism and memory access efficiency of the accelerator are l...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(China)
IPC IPC(8): G06F15/163G06F17/16G06N3/063
CPCG06F15/163G06F17/16G06N3/063Y02D10/00
Inventor 黄乐天明小满
Owner UNIV OF ELECTRONICS SCI & TECH OF CHINA
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products