Neural network processor based on efficient multiplex data stream, and design method

A neural network and design method technology, applied in neural learning methods, electrical digital data processing, biological neural network models, etc., can solve the problems of slow operation speed, operation speed bottleneck, and occupation of resources, etc., to improve energy efficiency, reduce On-chip data bandwidth, the effect of improving data sharing rate

Active Publication Date: 2017-08-22
INST OF COMPUTING TECHNOLOGY - CHINESE ACAD OF SCI
View PDF4 Cites 14 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0003] The deep network structure obtained by deep learning is an operation model, which contains a large number of data nodes, each data node is connected to other data nodes, and the connection relationship between each node is represented by weight. The mainstream neural network processing hardware includes general graphics processing However, with the continuous improvement of the complexity of the neural network, the problems of the neural network technology occupying a lot of resources, slow operation speed, and large energy consumption in the actual application process are becoming more and more prominent. , the applicability in mobile platforms or embedded platforms is not high, so there are serious energy efficiency problems and computing speed bottlenecks when this technology is applied in embedded devices or low-overhead data centers.

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Neural network processor based on efficient multiplex data stream, and design method
  • Neural network processor based on efficient multiplex data stream, and design method
  • Neural network processor based on efficient multiplex data stream, and design method

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0029] The purpose of the present invention is to provide a neural network processor and design method based on efficient multiplexing data streams. The processor adopts the time dimension-space dimension data stream and weight compression method in the existing neural network processor system, reducing the The on-chip data bandwidth is improved, the data sharing rate is improved, and invalid calculations are reduced, thereby improving the computing speed and operating energy efficiency of the neural network processor.

[0030] In order to achieve the above object, the neural network processor based on the efficient multiplexing data stream provided by the present invention includes:

[0031] At least one storage unit for storing operation instructions and operation data;

[0032] At least one calculation unit, used to perform neural network calculations; and a control unit, connected to the at least one storage unit and the at least one calculation unit, for obtaining the inf...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention puts forward a neural network processor based on efficient multiplex data stream, and a design method, and relates to the technical field of the hardware acceleration of neural network model calculation. The processor compares at least one storage unit, at least one calculation unit and a control unit, wherein the at least one storage unit is used for storing an operation instruction and arithmetic data; the at least one calculation unit is used for executing neural network calculation; and the control unit is connected with the at least one storage unit and the at least one calculation unit for obtaining the operation instruction stored by the at least one storage unit via the at least one storage unit, and analyzing the operation instruction to control the at least one calculation unit, wherein the arithmetic data adopts a form of the efficient multiplex data stream. By use of the processor, the efficient multiplex data stream is adopted in a neural network processing process, a weight and data only need to be loaded into one row of calculation unit in a calculation unit array each time, the bandwidth of data on chip is lowered, a data sharing rate is improved, and energy efficiency is improved.

Description

technical field [0001] The invention relates to the technical field of hardware acceleration for neural network model calculation, in particular to a neural network processor and a design method based on high-efficiency multiplexing data streams. Background technique [0002] With the continuous development of machine learning technology, deep neural network has become the best solution for cognition and recognition tasks, and has attracted widespread attention in the fields of recognition detection and computer vision, especially in the field of image recognition, deep neural network has reached or even surpassed Human recognition accuracy. [0003] The deep network structure obtained by deep learning is an operation model, which contains a large number of data nodes, each data node is connected to other data nodes, and the connection relationship between each node is represented by weight. The mainstream neural network processing hardware includes general-purpose graphics ...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
IPC IPC(8): G06F15/78G06F15/80G06N3/08
CPCG06F15/7807G06F15/8053G06N3/08Y02D10/00
Inventor 韩银和许浩博王颖
Owner INST OF COMPUTING TECHNOLOGY - CHINESE ACAD OF SCI
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products