Convolutional neural network hardware accelerator with novel convolution operation acceleration module

A technology of convolutional neural network and hardware accelerator, which is applied in the field of electronic information and deep learning, can solve the problems of data bus bit width, inability to consider each other, and large number of additions, etc., to achieve low data transmission bandwidth, reduce consumption, and reduce power consumption. consumption effect

Pending Publication Date: 2019-10-22
SOUTHEAST UNIV
View PDF2 Cites 2 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0004] Purpose of the invention: the purpose of the present invention is to solve the existing convolutional neural network operation method of the three ways still exist or the number of additions becomes more, or the low bandwidth of the data supply end makes the bit width of the data bus become wider to no avail, Either the registers added for caching data will consume a lot of D flip-flop resources, which is not good for the implementation of FPGA or ASIC. The three methods cannot take into account each other, and there is a problem of mutual conflict.

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Convolutional neural network hardware accelerator with novel convolution operation acceleration module
  • Convolutional neural network hardware accelerator with novel convolution operation acceleration module

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0033] The present invention will be further described below in combination with specific embodiments.

[0034] In the process of using the convolution kernel to perform convolution operations on the input feature map to obtain the output feature map, each convolution operation can obtain a pixel on the output feature map, and the convolution kernel slides to another area for the same convolution The operation can obtain additional output feature map pixels, and the convolution kernel can traverse the input feature map to obtain the entire output feature map. Therefore, the essence of calculating the output feature map is to repeat the same convolution operation multiple times, but the input is different. The present invention also follows this idea, as long as the hardware for performing a convolution operation is designed, and then reused in time, the purpose of calculating the complete output feature map can be achieved. Invent the process of performing a convolution operat...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention discloses a convolutional neural network hardware accelerator with a novel convolution operation acceleration module. The convolutional neural network hardware accelerator comprises an operation process management module; a parameter table module; a feature map caching module; a novel convolution operation acceleration module. The novel convolution operation acceleration module comprises: an input feature map pre-fetching module for reading input pixel points from a feature map cache; a product module which is used for multiplying the input pixel point by the convolution kernel weight; an output feature map access module which is responsible for reading an old intermediate result of the output pixel point from the feature map cache and writing a new intermediate result of theoutput pixel point back to the feature map cache; and an accumulation module which is used for completing the accumulation work of the operation result of the product module and the old intermediateresult of the output pixel point. According to the convolutional neural network hardware accelerator with the novel convolutional operation acceleration module, the convolutional operation process inthe convolutional neural network can be accelerated, and the convolutional neural network hardware accelerator has great application value in many mobile terminal devices with the characteristics of low data transmission bandwidth, low power consumption and few logic resources.

Description

technical field [0001] The invention relates to the technical fields of electronic information and deep learning, in particular to a convolution neural network hardware accelerator with a novel convolution operation acceleration module. Background technique [0002] In recent years, deep learning technology has developed rapidly, especially the convolutional neural network with a unique receptive field structure based on the working mechanism of cat brain visual cortex nerve cells has achieved great success in the field of vision applications, such as CNN in large image classification datasets On ImageNet, the recognition accuracy rate exceeds that of the human eye. However, this powerful algorithm encounters great obstacles in the process of moving towards practical applications. This is because many real-world applications where CNN can be useful are implemented on mobile devices with few computing resources, such as AI capabilities. Smartphones and smart security cameras...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
Patent Type & Authority Applications(China)
IPC IPC(8): G06N3/04G06N3/063G06K9/62
CPCG06N3/063G06N3/045G06F18/241
Inventor 张萌朱振宇贾贤飞柳飞扬
Owner SOUTHEAST UNIV
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products