Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Global rank perception neural network model compression method based on filter feature map

A neural network model and compression method technology, applied in the field of neural network model compression, global rank-aware neural network model compression based on filter feature map, can solve the problems of ill-conditioned filter, high labor cost, low training efficiency, etc. Expand the scope of application, reduce the difficulty of operation, and realize the effect of low complexity

Pending Publication Date: 2022-02-11
XIDIAN UNIV
View PDF0 Cites 6 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

Based on the assumption that each layer of the convolutional neural network is equally important, the weight of the unimportant filter can be minimized through the regularization operation, that is, the unified threshold can be applied to each convolutional layer of the network to achieve pruning, but the regularization operation exists The ill-conditioned problem of filter selection
He Y et al. in the document "He Y, Lin J, Liu Z, et al.Amc: Automl for model compression and acceleration on mobile devices. In: Proceedings of the European Conference on Computer Vision (ECCV), 2018.784-800." The proposed automatic model compression pruning model compression method AMC (Automated Model Compression), the filter pruning model compression method uses reinforcement learning to search for a convolutional neural network that meets the complexity specified by the user, but the learning time cost of the entire neural network is relatively high. Big
[0006] To sum up, most of the existing filter pruning model compression methods based on predefined structure or adaptive structure represent and sort the importance of local filters in the neural network model, but do not The importance of the filter is evaluated globally and uniformly. At the same time, it also faces the problems of low training efficiency and high labor cost when pruning general-purpose neural networks, and cannot achieve satisfactory results.

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Global rank perception neural network model compression method based on filter feature map
  • Global rank perception neural network model compression method based on filter feature map
  • Global rank perception neural network model compression method based on filter feature map

Examples

Experimental program
Comparison scheme
Effect test

Embodiment 1

[0031] Studies have shown that there is a lot of redundant information in the existing deep convolutional neural network model itself, including filter redundancy in the model convolution layer, and the high discriminative performance of the neural network model often depends on a few key parameters and Therefore, the filter pruning model compression method cuts out the filter redundancy in the neural network model, which is an effective model compression method, but the classification accuracy of the network model must be considered when the neural network model is pruned , reasoning speed, and the trade-offs between complexity.

[0032] Most of the existing filter pruning model compression methods are aimed at the importance representation, sorting and pruning of the network model for the local filters in the deep convolutional neural network, but there is no global sorting and unified evaluation for all the filters in the neural network model Filter importance, there is an ...

Embodiment 2

[0051] The global rank-aware neural network model compression method based on the filter feature map is the same as in embodiment 1, the average rank of the filter feature map in the pre-training model Ω is obtained as described in step (5), and the specific process includes the following steps:

[0052] (5.1) Get the feature maps of all filters in the neural network : In the training dataset X train Randomly select a sufficient number of images and corresponding labels in , and select the images as the input X of the pre-training model Ω in , the corresponding label is used as the output Y of the pre-training model Ω, and the feature map of all filters in the neural network is obtained using the hook Hook function

[0053]

[0054] Among them, C i Indicates the i-th convolutional layer in the pre-trained neural network model, Indicates that the hook Hook function acts on C i , n i-1 means C i The number of input channels, n i Indicates the number of filters, k ...

Embodiment 3

[0060] The global rank-aware neural network model compression method based on the filter feature map is the same as embodiment 1-2, using the preset pruning rate hyperparameter and pre-trained neural network model described in step (6), adaptive learning acquisition and preservation Parameter matrix α, γ, see Figure 5 , including the following steps:

[0061] (6.1) Parameter setting: set the parameters related to the genetic evolution algorithm, including the population size P, the maximum number of iterations E, the sample sampling size S, the mutation rate μ, the random step size σ, and the minimum constant , the number of fine-tuning iterations τ;

[0062] (6.2) Introduce related variables: introduce population queue, iterator and initialize, introduce coefficient parameter matrix α, bias parameter matrix γ, collectively referred to as parameter matrix;

[0063] (6.3) Parameter matrix update: the iterator is incremented by 1, the coefficient parameter matrix α and the b...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses a global rank perception neural network model compression method based on a filter feature map, and solves the problems of high labor cost, low pruning efficiency and poor stability of a filter pruning model compression method. The method comprises the steps of obtaining and preprocessing image data; constructing a universal convolutional neural network, and setting hyper-parameters; selecting a loss function and an optimization algorithm; training and storing the pre-training model; acquiring an average rank of the feature map; adaptively learning a parameter matrix; representing and sequencing the importance of the filters; pruning and storing the pre-training model; and finely adjusting the network model to realize the global rank perception neural network model compression based on the filter feature map. According to the method, one-time learning and multi-time pruning are adopted, the filters are globally sequenced, pruning is unified, and ill-conditioned selection does not exist. The pruning effect and stability are better, the pruning efficiency of large-scale different-complexity networks is high, and the adaptability of different edge devices is good. The method is used for computing and storing resource-limited edge equipment.

Description

technical field [0001] The invention belongs to the technical field of artificial intelligence, and mainly relates to neural network model compression, in particular to a global rank-aware neural network model compression method based on filter feature maps. It can be used for neural network deployment on edge devices with limited computing and storage resources. Background technique [0002] With the continuous development of artificial intelligence in the era of big data, the performance of deep convolutional neural networks is getting stronger and stronger, and has achieved great success in computer vision applications such as image classification, object detection, anomaly detection, semantic segmentation, and instance segmentation. However, along with it, the depth of the network model is getting deeper and larger, and the computing requirements are getting higher and higher. The deployment of most advanced convolutional neural networks in limited application environme...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
IPC IPC(8): G06V10/44G06V10/764G06V10/82G06N3/04G06N3/08
CPCG06N3/08G06N3/045G06F18/241
Inventor 谢卫莹卢天恩张鑫雷杰李云松马纪涛
Owner XIDIAN UNIV
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products