Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Convolutional neural network weight gradient optimization method based on data stream

A convolutional neural network and gradient optimization technology, which is applied in the field of convolutional neural network training, can solve problems such as time-consuming, storage resource consumption, and transmission time, and achieve obvious acceleration effects, speed up network training, and reduce transmission time.

Active Publication Date: 2021-04-09
TIANJIN UNIV
View PDF6 Cites 0 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

This scheme has two disadvantages: one is that it needs to read K×K times repeatedly, and the other is that it takes time to read data that does not participate in the calculation when calculating the corresponding gradient.
This method consumes a lot of storage resources and transmission time

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Convolutional neural network weight gradient optimization method based on data stream
  • Convolutional neural network weight gradient optimization method based on data stream
  • Convolutional neural network weight gradient optimization method based on data stream

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0027] The technical solutions of the present invention will be described in detail below in conjunction with the accompanying drawings and embodiments.

[0028] The weight gradient optimization calculation method of the convolutional neural network based on the data flow mode of the present invention optimizes the weight gradient. According to the formula, the gradient of the input image and the output image has a fixed difference (K-1). Such as image 3 As shown, it is a schematic diagram of a weight gradient optimization model of a convolutional neural network weight gradient optimization calculation method based on data flow in the present invention. The model slides the input image through a K×K sliding window, and changes the original unfixed convolution size to a fixed size K×K, which can support the calculation of weight gradients of any size. The specific calculation method is as follows: 1) Simultaneously calculate the input image covered by the first K×K window an...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses a convolutional neural network weight gradient optimization method based on a data stream, and provides a configurable data stream architecture design for convolutional neural network weight gradient optimization, so that convolution operations of different sizes in weight gradient calculation can be supported, and the degree of parallelism is K * K (convolution kernel size) times that of serial input. The training performance of the whole convolutional neural network is improved, and the problem that convolution operations of different sizes are difficult to realize in weight gradient calculation is solved. Compared with the prior art, the method has the advantages that (1) the acceleration effect is obvious: for weight gradient calculation, the degree of parallelism is improved by K * K compared with that of an original serial scheme, and the transmission time of data input is remarkably reduced, so that the purpose of accelerating the whole network training is achieved, and 1-1 / (K * K)% of input storage can be reduced compared with a general matrix multiplication scheme; and 2) the applicability and the universality are simultaneously met.

Description

technical field [0001] The invention belongs to the field of information technology and the field of convolutional neural network training hardware acceleration, and in particular relates to convolutional neural network training based on low power consumption and high performance. Background technique [0002] Convolutional Neural Network (CNN), as a feed-forward neural network, is widely used in various fields such as computer vision and natural language processing. With the increasing network scale and training data set of CNN, the training of CNN requires huge computing power, storage space and power consumption. The training of CNN includes the calculation of forward propagation and back propagation of the convolution module. At present, the academic and industrial circles have proposed many solutions for the hardware implementation of forward propagation, but there is a lack of hardware implementation for back propagation. Backpropagation includes the calculation of t...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(China)
IPC IPC(8): G06N3/08G06N3/04
CPCG06N3/082G06N3/045Y02D10/00
Inventor 刘强孟浩
Owner TIANJIN UNIV
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products