Check patentability & draft patents in minutes with Patsnap Eureka AI!

Lightweight convolutional neural network reconfigurable deployment method based on FPGA

A convolutional neural network, lightweight technology, applied in the field of artificial intelligence, can solve the problems of high computational and storage complexity of neural network algorithms, difficult to apply, and difficult to achieve high computing performance and energy efficiency.

Active Publication Date: 2020-11-13
BEIHANG UNIV
View PDF2 Cites 3 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0003]The current FPGA deployment scheme of convolutional neural network faces two challenges: First, the operating frequency of FPGA is usually 100-300MHz, which is much lower than that of CPU and GPU. At the same time, due to the limited on-chip resources of the FPGA, it is difficult to achieve high computing performance and energy efficiency only with a simple design
Secondly, due to the lack of development frameworks like Caffe and Tensorflow for developing FPGA-based applications, it is much more difficult to develop FPGA applications than CPU and GPU
Moreover, the computational and storage complexity of the neural network algorithm is very high, and it is an open problem to balance the model size and algorithm accuracy.
Large-scale networks are difficult to apply in application scenarios that have strict requirements on energy consumption and delay

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Lightweight convolutional neural network reconfigurable deployment method based on FPGA
  • Lightweight convolutional neural network reconfigurable deployment method based on FPGA
  • Lightweight convolutional neural network reconfigurable deployment method based on FPGA

Examples

Experimental program
Comparison scheme
Effect test

Embodiment

[0081] Such as figure 1 As shown, the FPGA-based lightweight convolutional neural network accelerator is designed and implemented on the Xilinx ZCU102 heterogeneous computing platform, and the convolution acceleration part is mainly deployed on the FPGA platform. The number of available DSPs on the Xilinx ZCU102 platform is 2520, and the number of available 36K BRAMs is 912. At present, the high-end FPGA platform integrates hard-core DSP for high-speed computing. DSP refers to the on-chip computing resources of the FPGA. The number of available DSPs refers to the unoccupied amount of the hard-core DSP on the FPGA platform, which is used to evaluate the FPGA on-chip computing. Capacity can be increased.

[0082] In this embodiment, the Yolo-Tiny backbone network is used as the target network, which includes 9 convolutional layers, 9 ReLU layers and 6 maximum pooling layers. The input feature map size of the first layer is 416×416×3, The output feature map size of the last lay...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention provides a lightweight convolutional neural network reconfigurable deployment method based on an FPGA. During deployment, adaptation to FPGA platforms with different resource conditionsis realized by configuring parameters such as the column number of convolution windows and the number of output channels; before operation, adaptation to different network layers is realized by configuring convolution kernel sizes, pooling step sizes, input and output feature map sizes and input and output channel numbers of different convolution layers; and constructing an FPGA-based lightweightconvolutional neural network accelerator to realize network reconfigurable deployment, wherein the lightweight convolutional neural network accelerator comprises an off-chip DDR data interaction unit,an on-chip storage buffer processing unit, a computing engine and a central control unit. According to the method, different implementation schemes can be selected for deployment for different FPGA platforms, so that existing resources of the FPGA platforms are utilized to the maximum extent, the characteristics of high parallelism, high throughput rate and low power consumption of the FPGA platforms are brought into full play, and convolutional neural network reasoning acceleration is realized efficiently and quickly under the condition of relatively low power consumption.

Description

technical field [0001] The present invention relates to the technical field of artificial intelligence, and more specifically relates to an FPGA-based lightweight convolutional neural network reconfigurable deployment method. Background technique [0002] At present, because the CPU platform cannot provide enough computing power, the GPU has become the preferred neural network processing platform due to its powerful computing performance. However, the GPU consumes a lot of power, and it is difficult to deploy it on an embedded platform that requires high power consumption. The FPGA-based neural network reasoning accelerator has become a new research hotspot, and is expected to surpass the GPU platform in terms of computing speed and energy efficiency. [0003] The current FPGA deployment scheme for convolutional neural networks faces two challenges: First, the operating frequency of FPGA is usually 100-300MHz, which is much lower than that of CPU and GPU. Designs are diffi...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(China)
IPC IPC(8): G06N3/04G06N3/063
CPCG06N3/063G06N3/045Y02D10/00
Inventor 李波刘晓戬姜宏旭张永华
Owner BEIHANG UNIV
Features
  • R&D
  • Intellectual Property
  • Life Sciences
  • Materials
  • Tech Scout
Why Patsnap Eureka
  • Unparalleled Data Quality
  • Higher Quality Content
  • 60% Fewer Hallucinations
Social media
Patsnap Eureka Blog
Learn More