Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

A Dynamically Reconfigurable Convolutional Neural Network Accelerator Architecture for the Internet of Things

A convolutional neural network and Internet of Things technology, applied in the field of dynamic reconfigurable convolutional neural network accelerator architecture, can solve the problems of energy efficiency (low performance/power consumption, inability to apply smart mobile terminals, high power consumption, etc.) The effects of external memory access, simple network structure, and low power consumption

Active Publication Date: 2022-02-22
XI AN JIAOTONG UNIV
View PDF6 Cites 0 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

However, the existing hardware implementation has high power consumption and low energy efficiency (performance / power consumption), and cannot be applied to smart mobile terminals, such as smartphones, wearable devices or self-driving cars. Wait
In this context, reconfigurable processors have been proven to be a form of parallel computing architecture with both high flexibility and high energy efficiency. Improving processing performance while using a dedicated processor is one of the solutions to the limitations of the further development of multi-core CPU and FPGA technology, and may become one of the solutions for realizing high-performance deep learning SoC in the future

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • A Dynamically Reconfigurable Convolutional Neural Network Accelerator Architecture for the Internet of Things
  • A Dynamically Reconfigurable Convolutional Neural Network Accelerator Architecture for the Internet of Things
  • A Dynamically Reconfigurable Convolutional Neural Network Accelerator Architecture for the Internet of Things

Examples

Experimental program
Comparison scheme
Effect test

Embodiment

[0090] Regarding the speed index, the superiority of the present invention comes from the design of the processing unit array and cache architecture. First, the processing unit adopts the Winograd convolution acceleration algorithm. For example, for a convolution operation with a size of 5*5 input data, a convolution kernel size of 3*3, and a step size of 1, traditional convolution needs to introduce 81 multiplication operations, while this It is published that each processing unit only needs to introduce 25 multiplications. In addition, the processing unit array is in the convolutional network, and the input channel and the output channel are processed in a certain degree of parallelism, which makes the convolution operation faster. On the other hand, the cache architecture has two working modes. In the on-chip working mode, the data generated by the middle layer of the convolutional neural network does not need to be stored off-chip, but can be directly sent to the next layer...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The present invention is a dynamic reconfigurable convolutional neural network accelerator architecture for the Internet of Things field, including a cache architecture, etc. The cache architecture is used to store data from external storage or data generated during the calculation process, and organize and arrange them , is transmitted to the processing unit array in a data structure for calculation; the processing unit array is used to receive data from the cache architecture, and is stored in the cache architecture after convolution operation processing; the calculation module is used to receive data from the processing unit array, select Perform three operations of pooling, normalization, or activation functions, and store the output data in the cache architecture; the controller is used to send commands to the cache architecture, processing unit array, and computing module, and is designed with an external interface for communicating with the outside system to communicate. The present invention improves the performance of the convolutional neural network accelerator and reduces power consumption by designing a processing unit array with high parallelism and high utilization rate and a cache architecture that can increase the data multiplexing rate.

Description

technical field [0001] The invention belongs to the field of neural network accelerators, and in particular relates to a dynamic reconfigurable convolutional neural network accelerator architecture for the Internet of Things field. Background technique [0002] Artificial intelligence is one of the current popular computer sciences. As the main way to realize artificial intelligence, deep learning has also achieved profound development. As the number of network layers and the number of neurons in each layer increase, the computational complexity of the model will increase with As the size of the network increases, it grows exponentially. Therefore, the bottleneck of learning speed and running speed of deep learning algorithms is increasingly dependent on large-scale computing platforms such as cloud computing. For the hardware acceleration of deep learning algorithms, there are currently three types of implementations—multi-core CPUs, GPUs, and FPGAs. Their common feature i...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Patents(China)
IPC IPC(8): G06N3/04G06N3/063G06N3/08
CPCG06N3/063G06N3/08G06N3/045
Inventor 杨晨王逸洲王小力耿莉
Owner XI AN JIAOTONG UNIV
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products