Neural network structured sparse method based on incremental regularization

A technology of network structure and neural network model, applied in the direction of neural learning method, biological neural network model, neural architecture, etc., can solve the problems of weak expressiveness and performance loss of CNN, and achieve obvious effect, great flexibility, and good pruning effect of effect

Inactive Publication Date: 2019-09-03
ZHEJIANG UNIV
View PDF0 Cites 8 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0004] In order to solve the problem of performance loss caused by using a large and constant regularization factor in the above-mentioned traditional regularization method to obtain a better model pruning effect, the present invention provides a neural network structure based on incremental regularization sparse method
At the same time, when pruning, this method first initializes all regularization factors to 0, and then gradually increases the regularization factors based on the relative importance of each weight group, which solves the traditional algorithm due to CNN The problem of weak expressiveness and difficult to bear large penalties at the beginning of pruning, thereby reducing the performance loss of the model

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Neural network structured sparse method based on incremental regularization
  • Neural network structured sparse method based on incremental regularization

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0029] In conjunction with the following implementation examples, the present invention is further described in detail. However, the neural network structured column sparse algorithm proposed by the present invention is not limited to this implementation method.

[0030] The present invention provides a neural network structured sparse method based on incremental regularization. The overall network model sparse process is as follows: figure 1 shown.

[0031] (1) Preparation work

[0032] For the neural network model to be sparse, prepare the training data set, network structure configuration file, and training process configuration file. The used data set, network structure configuration, and training process configuration are all consistent with the original training method; in ResNet-50 In the neural network structured column sparse experiment, the dataset used is ImageNet-2012, and the network structure configuration and other files used are the files used by the original...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention discloses a neural network structured sparse method based on incremental regularization. The method comprise the steps of when the branches of a neural network are pruned, according to the relative importance of each weight group, distributing different regularization increments to different weight groups step by step, then updating the regularization factors of the weight groups iteratively and continuously, and when the regularization factor of a certain weight group reaches a specified regularization upper limit, deleting the corresponding weight in the network permanently toincrease the structured sparsity of the network model; and when the sparsity of a certain layer reaches a preset sparsity rate, automatically stopping pruning of the layer until the pruning of all thelayers is completed; and finally, retraining the whole network to callback the accuracy, and when the accuracy of the model does not rise any more, stopping retraining to obtain a sparse model. According to the method, a large deep learning model can be deployed on the mobile and embedded equipment, a remarkable actual acceleration effect is obtained, and the application of a deep learning algorithm on a mobile terminal is promoted.

Description

technical field [0001] The invention relates to the fields of neural calculation, pattern recognition and computer vision, in particular to a method for pruning the weight of a neural network to enhance its generalization ability, reduce storage capacity and speed up operation. Background technique [0002] In recent years, deep convolutional neural networks (CNN) have achieved remarkable success in computer vision tasks by using large amounts of data to learn large-scale networks. However, CNNs usually incur heavy computation and storage consumption, hindering their deployment on mobile and embedded devices. To address this issue, much research work has focused on compressing the size of CNNs. As a more effective method, parameter pruning is used to compress and accelerate CNN models, and its purpose is to eliminate redundant model parameters within a tolerable performance loss range. However, the general parameter pruning method may generate unstructured random connectio...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
Patent Type & Authority Applications(China)
IPC IPC(8): G06N3/04G06N3/08
CPCG06N3/082G06N3/045
Inventor 胡浩基李翔王欢
Owner ZHEJIANG UNIV
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products