Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Multi-layer feedforward neural network parallel accelerator

A feed-forward neural network and accelerator technology, applied in the field of multi-layer feed-forward neural network parallel accelerators, can solve the problems of data migration and calculation that cannot achieve higher efficiency, transistors cannot function, and poor scalability, etc., to achieve resource consumption Fewer, adjustable parallelism, good scalability

Active Publication Date: 2018-10-26
NANJING UNIV
View PDF3 Cites 4 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

However, both GPU and CPU are general-purpose processors, and both need to perform the process of fetching instructions, decoding instructions, and executing instructions. In this way, the processing of the underlying IO is shielded, so that the software and hardware are decoupled, but it brings data movement and Computing cannot achieve higher efficiency
The difference in the energy consumption ratio between the GPU and the CPU is mainly due to the fact that most of the transistors in the CPU are used in the Cache and the control logic unit. Therefore, compared with the GPU, the CPU has less redundancy for algorithms that are computationally intensive and have low computational complexity. The remaining transistors can't play a role, and the energy consumption is lower than that of CPU than GPU
The above two implementation methods consume a lot of energy and have poor scalability. How to make the multi-layer feedforward neural network run faster, save more energy, and have better scalability has become a hot issue.

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Multi-layer feedforward neural network parallel accelerator
  • Multi-layer feedforward neural network parallel accelerator
  • Multi-layer feedforward neural network parallel accelerator

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0027] The present invention will be described in detail below in conjunction with the accompanying drawings and specific implementation cases.

[0028] Such as figure 1 As shown, a typical neuron such as figure 1 Shown: Receive input signals from n other neurons, these input signals are transmitted through weighted connections, the total input value received by the neuron will be compared with the neuron threshold, and then processed by the "activation function" Produce neuron output.

[0029] The multi-layer feed-forward neural network parallel accelerator of this embodiment is mainly composed of a main control module, a data division module, a coefficient address generation module, an operand address generation module, a neuron address generation module, a storage control module and a storage module, see figure 2 .

[0030] Among them, the main control module receives the system start signal, calls the data division module to allocate the calculation of hidden layer neu...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

A multi-layer feedforward neural network parallel accelerator provided by the invention comprises: a main control module, which controls the entire calculation flow; a coefficient address generation module, which generates an address of the coefficient, outputs the address of the coefficient, receives coefficient data of the storage control module, and outputs the coefficient data after splitting;an operand address generation module, which generates and outputs the storage address of the operand, receives the operand data of the storage control module, and outputs the data after splitting; afeedforward network calculation module, which receives the split coefficient data and the operand data and contains a plurality of parallel computing units; and a neuron address generating module, which receives the neuron data calculated by the feedforward network computing module, generates the storage address of the neuron and neuron data, and outputs the storage address of the neuron and the neuron data. The invention has the beneficial effects of good scalability, adjustable parallelism, high acceleration ratio, support for pipeline operation, and low resource consumption.

Description

technical field [0001] The invention belongs to the field of hardware acceleration, in particular to a multilayer feedforward neural network parallel accelerator. Background technique [0002] Artificial intelligence algorithms make people's lives more and more convenient, but with the continuous change of application requirements, the algorithm complexity is also getting higher and higher. With the increase of algorithm complexity, the requirements of artificial intelligence algorithms for hardware power consumption and speed are also increasing. A neural network is an extensively parallel interconnected network of adaptive, simple units organized to simulate the interaction of biological nervous systems with real-world objects. Multi-layer feed-forward neural network has been more and more widely used in operand retrieval, machine vision, security monitoring and other fields. Note that the input of the neuron in the jth hidden layer of the multi-layer feedforward neural ...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
IPC IPC(8): G06N3/063
CPCG06N3/063
Inventor 李丽李宏炜樊朝煜潘红兵何书专陈沁雨
Owner NANJING UNIV
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products