Highly flexible configurable data post-processor for deep neural network

A technology of deep neural network and post-processor, applied in the field of data post-processor, can solve the problems of low hardware reuse rate, consumption of hardware resources, unsatisfactory cost, etc., to meet design needs, flexible use, and reduce hardware cost Effect

Pending Publication Date: 2020-10-16
上海赛昉科技有限公司
View PDF0 Cites 15 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0003] Although the dedicated hardware architecture can provide higher computing speed, each layer of the network will be implemented by a dedicated hardware circuit, so there will be a dedicated MAC matrix and data post-processing module. Such an architecture has a low hardware reuse rate and consumes More hardware resources cannot meet the needs of some cost-sensitive usage scenarios

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Highly flexible configurable data post-processor for deep neural network
  • Highly flexible configurable data post-processor for deep neural network
  • Highly flexible configurable data post-processor for deep neural network

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0028] A highly configurable data post-processor for deep neural networks. The configuration of the processor is divided into pre-run (AOT) configuration and runtime (RT) configuration. The AOT configuration works in the data post-processor The hardware implementation stage is used to obtain the data post-processor required by the user, such as figure 1 As shown, in order to realize the AOT configuration, the present invention designs a dedicated hardware generator for the hardware implementation stage of the data post-processor. The hardware generator is divided into a configuration layer and an implementation layer. In the configuration layer, data post-processing The configuration information implemented by the hardware of the processor, including the data bit width of the data path in the data post-processor, the number of registers of key data processing nodes, etc., and the configuration layer has an interface that can be operated, and the user can fill in it as needed F...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention discloses a highly flexible configurable data postprocessor for a deep neural network, the configuration mode of the processor is divided into pre-operation (AOT) configuration and runtime (RT) configuration, and the AOT configuration mode works in the hardware implementation stage of the data postprocessor and is used for obtaining the data postprocessor required by a user; the RT configuration mode works in the operation stage of the deep neural network acceleration engine, and at the moment, hardware implementation of the data postprocessor is completed. According to the invention, hardware resources which can be used by the data postprocessor are determined through AOT configuration; a data processing path of the data postprocessor during operation is determined through RT configuration; a more flexible use mode is provided for a user of the data postprocessor, balance among power consumption, performance and cost is facilitated, meanwhile, the situation that the special data postprocessor is designed due to the fact that the special data postprocessor needs to be matched with different network layers is avoided, hardware cost is reduced, and the design requirement of a cost-sensitive deep neural acceleration engine is met.

Description

technical field [0001] The invention relates to the technical field of deep neural network processors, in particular to a highly flexibly configurable data post-processor for deep neural networks. Background technique [0002] Deep neural network is a widely used machine learning algorithm. In deep neural network, it usually includes standard convolution layer, pooling layer, fully connected layer and activation layer. At present, the common dedicated deep neural network accelerators widely use dedicated hardware circuits to support the operation of each layer of the neural network. This dedicated hardware circuit usually consists of a MAC matrix and a data post-processing module. The data post-processing module will perform different processing on the MAC matrix output data for different network layers. For example, the convolution layer will perform cumulative calculations. In addition, the data post-processor can also support additional operations required for data block...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
Patent Type & Authority Applications(China)
IPC IPC(8): G06N3/063G06N3/04G06F9/50
CPCG06N3/063G06F9/5027G06N3/045Y02D10/00
Inventor 李思彧伍骏王维
Owner 上海赛昉科技有限公司
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products