Hardware acceleration system for LSTM (Long Short Term Memory) network model

A network model, hardware acceleration technology, applied in biological neural network model, climate sustainability, neural architecture, etc., can solve the problems of lack of research results and general optimization effect at the computing level.

Active Publication Date: 2021-07-30
HUAZHONG UNIV OF SCI & TECH
View PDF5 Cites 2 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

Although this method realizes the acceleration of the LSTM model system, it mainly optimizes the bottleneck of storage resource consumption, and the optimization effect at the computing level is average.
[0006] In summary, although the research on LSTM model acceleration tasks for hardware platforms is of great significance, the research results in this area are still relatively scarce, and an LSTM network acceleration design with good versatility and excellent acceleration effects is urgently needed

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Hardware acceleration system for LSTM (Long Short Term Memory) network model
  • Hardware acceleration system for LSTM (Long Short Term Memory) network model
  • Hardware acceleration system for LSTM (Long Short Term Memory) network model

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0072] In order to make the object, technical solution and advantages of the present invention clearer, the present invention will be further described in detail below in conjunction with the accompanying drawings and embodiments. It should be understood that the specific embodiments described here are only used to explain the present invention, not to limit the present invention. In addition, the technical features involved in the various embodiments of the present invention described below can be combined with each other as long as they do not constitute a conflict with each other.

[0073] Such as figure 1 Shown is an LSTM network-oriented hardware acceleration architecture disclosed in the embodiment of the present invention, wherein the "off-chip memory" is an external off-chip storage device, and the "on-chip processing unit" is the main part of the architecture in this application, mainly including :

[0074] Network reasoning computing core: As a computing accelerato...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention discloses a hardware acceleration system for an LSTM network model, and belongs to the technical field of deep learning hardware acceleration. The invention discloses a hardware acceleration system for a deep learning long short-term memory (LSTM) network model. The hardware acceleration system comprises a network reasoning calculation core and a network data storage core. The network reasoning calculation core is used as a calculation accelerator of an LSTM network model, calculation units are deployed according to the network model, and calculation acceleration of the calculation units such as convolution operation, matrix dot multiplication, matrix addition and an activation function is achieved; and the network data storage core serves as a data cache and interaction controller of an LSTM network model, deploys an on-chip cache unit according to the network model, and realizes a data interaction link between the computing core and an off-chip memory. According to the invention, the calculation parallelism of the LSTM network model is improved, the processing delay is reduced, the memory access time is shortened, and the memory access efficiency is improved.

Description

technical field [0001] The invention belongs to the field of deep learning hardware acceleration, and more specifically relates to a hardware acceleration system oriented to an LSTM network model. Background technique [0002] Long Short-Term Memory (LSTM), as a variant of deep learning Recurrent Neural Network (RNN), is widely used in sequence model processing tasks such as speech recognition, natural language processing, and image compression. LSTM effectively solves the problem of gradient explosion and gradient disappearance in the RNN training process by introducing a gating mechanism and a state value for storing long-term and short-term historical information, and relatively greatly increases its computational complexity and space complexity. Its intensive calculation and memory access limit its application on the embedded hardware platform with limited resources. Therefore, it is a very meaningful research to design and accelerate the optimization of the LSTM model f...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
Patent Type & Authority Applications(China)
IPC IPC(8): G06N3/04G06N3/06G06N3/08
CPCG06N3/08G06N3/06G06N3/044G06N3/045Y02D10/00
Inventor 钟胜王煜颜露新邹旭陈立群徐文辉张思宇颜章
Owner HUAZHONG UNIV OF SCI & TECH
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products