Supercharge Your Innovation With Domain-Expert AI Agents!

A convolutional network accelerator, configuration method, and computer-readable storage medium

A technology of convolutional network and configuration method, which is applied in the direction of data exchange network, digital transmission system, electrical components, etc., can solve the problems of not making full use of FPGA resources, not having a general implementation plan, and not achieving time-division multiplexing, etc. Achieve flexible configurability, balance between time and accelerator operation time, and increase bandwidth

Active Publication Date: 2021-09-03
HUAZHONG UNIV OF SCI & TECH
View PDF10 Cites 0 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0005] To sum up, the existing problems in the existing technology in many convolutional network hardware acceleration schemes implemented on FPGA are: (1) they are all aimed at a specific network structure model,
[0006] (2) All the network layers of the model are implemented on the chip. This method can only be used for some smaller networks, or the FPGA resources are not fully utilized.
[0007] (3) Time-sharing multiplexing is not achieved, and the degree of parallelism is low
[0008] Difficulty in solving the above technical problems: Among the public solutions for implementing convolutional neural networks on FPGA, there is no good general-purpose implementation solution, and at the same time, resource consumption needs to be considered while achieving high parallelism

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • A convolutional network accelerator, configuration method, and computer-readable storage medium
  • A convolutional network accelerator, configuration method, and computer-readable storage medium
  • A convolutional network accelerator, configuration method, and computer-readable storage medium

Examples

Experimental program
Comparison scheme
Effect test

Embodiment

[0068] The FPGA chip-based convolutional network accelerator provided by the embodiment of the present invention is based on the basic structure of a single-layer convolutional network, that is, the structure of convolutional layer + pooling layer + activation layer + batch normalization operation layer. For the number of layers of the overall network model, the executed forward network layer obtains the configuration parameters of the current layer, such as the size of the input and output feature map (length, width, number of channels), and the size of the convolution kernel (length, width, number of channels). , step size of convolution and pooling operations, etc., and load feature maps and weight parameters in batches from DDR (double data rate off-chip memory) through configuration parameters. At the same time, the acceleration kernel of the convolutional layer can also configure the degree of parallelism according to the configuration parameters.

[0069] The present in...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention belongs to the technical field of hardware acceleration of convolutional networks, and discloses a convolutional network accelerator, a configuration method, and a computer-readable storage medium, and judges the number of layers of the currently executed forward network layer in the overall network model by marking, and obtains The currently executed forward network layer configures parameters, and loads the feature map and weight parameters from the DDR through the configuration parameters; at the same time, the acceleration core of the convolution layer also configures the degree of parallelism according to the obtained executed forward network layer configuration parameters. The invention changes the network layer structure by configuring parameters, so that only one layer structure can be used when the network FPGA is deployed, which not only achieves flexible configurability, but also achieves the effect of saving and fully utilizing FPGA on-chip resources. The method of combining multiple RAMs into an overall cache area improves the bandwidth of data input and output. At the same time, the ping-pong operation is used to allow the loading of feature maps and weight parameters and the operation of the accelerator to achieve pipeline work.

Description

technical field [0001] The invention belongs to the technical field of hardware acceleration of convolutional networks, and in particular relates to a convolutional network accelerator, a configuration method and a computer-readable storage medium. Background technique [0002] At present, with the development of deep learning technology, convolutional neural network is more and more widely used in computer vision, such as target detection and recognition, tracking, semantic segmentation, speech recognition and natural language processing. Its outstanding data fitting performance And the versatility of the model makes the application of convolutional neural network in various complex scene fields replace the original traditional modeling method and become the benchmark in this field. But at the same time, the powerful data fitting ability is at the cost of a huge amount of data and calculation. For example, the model size of AlexNet is 233MB, and the calculation amount is 0....

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Patents(China)
IPC IPC(8): H04L12/24
CPCH04L41/082H04L41/14H04L41/142H04L41/145
Inventor 钟胜卢金仪颜露新王建辉徐文辉颜章唐维伟李志敏
Owner HUAZHONG UNIV OF SCI & TECH
Features
  • R&D
  • Intellectual Property
  • Life Sciences
  • Materials
  • Tech Scout
Why Patsnap Eureka
  • Unparalleled Data Quality
  • Higher Quality Content
  • 60% Fewer Hallucinations
Social media
Patsnap Eureka Blog
Learn More