Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

FPGA (Field Programmable Gate Array) platform-oriented recurrent neural network algorithm optimization method

A cyclic neural network and neural network algorithm technology, applied in the field of cyclic neural network algorithm optimization, can solve the problems of efficiency, poor computing speed, insufficient mobile performance, and high power consumption, so as to reduce computing delay and power consumption, and improve computing efficiency. and resource utilization

Pending Publication Date: 2022-08-02
SHENYANG AIRCRAFT DESIGN INST AVIATION IND CORP OF CHINA
View PDF0 Cites 0 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

But this also brings challenges to the computing power of the computer
At present, the training of deep learning models mostly adopts CPU and GPU solutions. Because of the structural characteristics of CPU, the efficiency and calculation speed of using CPU to train deep learning models are not as good as those of dedicated graphics processors GPU.
However, GPU-based accelerated training models have the disadvantages of high power consumption and insufficient mobile performance, and are difficult to use in application scenarios that require low power consumption, strong real-time performance, and pursuit of mobile performance.

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • FPGA (Field Programmable Gate Array) platform-oriented recurrent neural network algorithm optimization method
  • FPGA (Field Programmable Gate Array) platform-oriented recurrent neural network algorithm optimization method

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0013] In order to make the objectives, technical solutions and advantages of the implementation of the present application clearer, the technical solutions in the embodiments of the present application will be described in more detail below with reference to the accompanying drawings in the embodiments of the present application. Throughout the drawings, the same or similar reference numbers refer to the same or similar elements or elements having the same or similar functions. The described embodiments are some, but not all, embodiments of the present application. The embodiments described below with reference to the accompanying drawings are exemplary, and are intended to be used to explain the present application, but should not be construed as a limitation to the present application. Based on the embodiments in the present application, all other embodiments obtained by those of ordinary skill in the art without creative work fall within the protection scope of the present...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention belongs to the technical field of data processing, and particularly relates to an FPGA platform-oriented recurrent neural network algorithm optimization method. The method comprises the following steps of: storage optimization: storing intermediate variables in a neural network algorithm by adopting a distributed RAM (Random Access Memory) constructed by an LUT (Loop Under Test) in an FPGA (Field Programmable Gate Array), and storing parameters with storage resource requirements exceeding a threshold value in the neural network algorithm by adopting a BRAM (Block Random Access Memory) in the FPGA; and algorithm optimization: for vector multiplication and addition operation in the forward calculation process of the neural network algorithm, carrying out multi-loop operation on the outer layer of the vector multiplication and addition operation, and carrying out pipeline optimization on the calculation process. According to the invention, the calculation efficiency and the resource utilization rate of the recurrent neural network are improved, and the calculation time delay and power consumption are reduced.

Description

technical field [0001] The present application belongs to the technical field of data processing, and particularly relates to an FPGA platform-oriented cyclic neural network algorithm optimization method. Background technique [0002] Machine learning and deep learning are applied in different scenarios, such as text classification, object recognition, speech recognition, and autonomous driving. Even in some specific application scenarios, the performance of artificial intelligence is better than human judgment, such as Google's alpha go robot trained by reinforcement learning to beat the world champion in the game of Go. Data-driven is an important feature of deep learning, and a large amount of high-quality data can train models with higher accuracy and stronger capabilities. But this also brings challenges to the computing power of computers. At present, most of the training deep learning models use CPU and GPU solutions. Because of the structural characteristics of CPU...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
IPC IPC(8): G06N3/063G06N3/08G06F30/34G06F30/27
CPCG06N3/063G06N3/08G06F30/34G06F30/27Y02D10/00
Inventor 赵英策孙智孝罗庆
Owner SHENYANG AIRCRAFT DESIGN INST AVIATION IND CORP OF CHINA
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products