Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Multi-mechanism mixed recurrent neural network model compression method

A recursive neural network and model technology, which is applied in the field of recurrent neural network model compression, can solve the problem that the recurrent neural network cannot adapt to the storage resources and computing power of embedded systems, and achieve the effect of wide application prospects

Inactive Publication Date: 2018-01-30
南京风兴科技有限公司
View PDF0 Cites 9 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0006] The technical problem to be solved by the present invention is to propose a variety of efficient model compression mechanisms and hybrid methods for recurrent neural networks that cannot adapt to the storage resources and computing capabilities of embedded systems, so that the application of recurrent neural networks in embedded systems becomes possible

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Multi-mechanism mixed recurrent neural network model compression method
  • Multi-mechanism mixed recurrent neural network model compression method
  • Multi-mechanism mixed recurrent neural network model compression method

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0022] Embodiments of the present invention are described in detail below. Since the recurrent neural network includes many variants, this embodiment will take the most basic recurrent neural network as an example, which is intended to explain the present invention, but should not be construed as a limitation to the present invention. The implementation process of the rest of the recurrent neural network variants is basically the same as this embodiment.

[0023] A typical formulation of a recurrent neural network with m input nodes and n output nodes is defined as:

[0024] h t =f(Wx t +Uh t-1 +b), (1)

[0025] Among them, h t ∈ R n×1 is the hidden state of the recurrent neural network at time t; x t ∈ R m×1 is the input vector at time t; W∈R n×m , U∈R n×n , b∈R n×1 is the model parameter of the recurrent neural network, b is the bias item, W and U are the parameter matrix; f is a nonlinear function, and the common ones are σ and tanh:

[0026]

[0027] Such as...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses a multi-mechanism mixed recurrent neural network model compression method. The multi-mechanism mixed recurrent neural network model compression method comprises A, carrying outcirculant matrix restriction: restricting a part of parameter matrixes in the recurrent neural network into circulant matrixes, and updating a backward gradient propagation algorithm to enable the network to carry out batch training of the circulant matrixes, B, carrying out forward activation function approximation: replacing a non-linear activation function with a hardware-friendly linear function during the forward operation process, and keeping the backward gradient updating process unchanged; C, carrying out hybrid quantization: employing different quantification mechanisms for differentparameters according to the error tolerance difference between different parameters in the recurrent neural network; and D, employing a secondary training mechanism: dividing network model training into two phases including initial training and repeated training. Each phase places particular emphasis on a different model compression method, interaction between different model compression methodsis well avoided, and precision loss brought by the model compression method is reduced to the maximum extent. According to the invention, a plurality of model compression mechanisms are employed to compress the recurrent neural network model, model parameters can be greatly reduced, and the multi-mechanism mixed recurrent neural network model compression method is suitable for a memory-limited andlow-delay embedded system needing to use the recurrent neural network, and has good innovativeness and a good application prospect.

Description

technical field [0001] The invention relates to the technical field of computer and electronic information, in particular to a multi-mechanism mixed recursive neural network model compression method. Background technique [0002] Recurrent neural network has powerful nonlinear fitting ability, and its natural recursive structure is very suitable for modeling sequence data, such as text, speech and video. At present, the recurrent neural network model has achieved an effect or accuracy close to or even surpassing that of humans in the field of natural language processing, especially in speech recognition and machine translation; It also has broad application prospects. These technologies are necessary to realize intelligent human-computer interaction, but there are many problems in running recurrent neural network models on embedded devices. On the one hand, the recurrent neural network model needs to store a large number of parameters, and the amount of calculation is huge...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
IPC IPC(8): G06N3/08
Inventor 王中风王智生林军
Owner 南京风兴科技有限公司
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products