Supercharge Your Innovation With Domain-Expert AI Agents!

Method for batch normalization layer pruning in deep neural networks

一种深度神经网络、模型的技术,应用在深度神经网络领域,能够解决占用大运算资源、庞大运算能力、困难等问题,达到规模降低、确保准确性、减少硬件规格的效果

Inactive Publication Date: 2019-12-10
KNERON TAIWAN CO LTD
View PDF0 Cites 0 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

However, in the model inference stage, these batch normalization layers will consume a lot of computing resources and cause the device to require huge computing power to handle the training of the neural network.
Therefore, it is still difficult to import deep neural networks with batch normalization layers to all types of hardware, such as central processing units, graphics processing units, digital signal processors, etc.

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Method for batch normalization layer pruning in deep neural networks
  • Method for batch normalization layer pruning in deep neural networks
  • Method for batch normalization layer pruning in deep neural networks

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0015] In order to solve the problem of implementing deep neural network training with batch normalization layer in devices with limited computing resources, we propose a new batch normalization layer pruning technique, which combines linear operation layer pruning with (linear operation layer) any batch normalization layer connected to losslessly compress deep neural network models. Linear operation layers include but are not limited to convolution layers, dense layers, depthwise convolution layers and group convolution layers. In addition, the batch normalization layer pruning technique does not change the structure of other layers in the neural network model, so the batch normalization layer pruning technique can be directly implemented into the existing neural network model platform.

[0016] Before disclosing the details of the batch normalization layer pruning technique, the main claims (but not all claims) of this patent application are summarized here.

[0017] Embodi...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

A method of pruning a batch normalization layer from a pre-trained deep neural network model is proposed. The pre-trained deep neural network model is inputted as a candidate model. The candidate model is pruned by removing the at least one batch normalization layer from the candidate model to form a pruned candidate model only when the at least one batch normalization layer is connected to and adjacent to a corresponding linear operation layer. Weights of the corresponding linear operation layer are adjusted to compensate for the removal of the at least one batch normalization. The pruned candidate model is then output and utilized for inference. In the method, by pruning batch normalization layers connecting with a linear operation layer in a pre-trained DNN, the size of the DNN is reduced and implementation requirements are also reduced. Model inference can be achieved with much increased speed and much decreased computational requirements while guaranteeing accuracy during inference.

Description

technical field [0001] The present invention relates to a deep neural network with at least one batch normalization layer, and more particularly to a method for pruning a batch normalization layer from a pre-trained deep neural network model. Background technique [0002] Large-scale deep neural network (deep neural network, DNN) has significant functions in the fields of machine vision, image recognition and speech processing. However, these modern deep neural networks usually contain many layers, have very large models, and require high computational intensity. These characteristics make it very difficult to use deep neural networks on end-user devices (e.g., mobile phones) with low memory and low computing power. Modern deep neural networks such as ResNet, MobileNet, GoogLeNet, and Xception contain multiple batch normalization layers. These batch normalization layers are usually placed before or after convolutional, dense or deep convolutional layers to help train the n...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(China)
IPC IPC(8): G06N3/04G06N3/08
CPCG06N3/08G06N3/045G06N3/082G06N3/04
Inventor 谢必克苏俊杰张博栋刘峻诚
Owner KNERON TAIWAN CO LTD
Features
  • R&D
  • Intellectual Property
  • Life Sciences
  • Materials
  • Tech Scout
Why Patsnap Eureka
  • Unparalleled Data Quality
  • Higher Quality Content
  • 60% Fewer Hallucinations
Social media
Patsnap Eureka Blog
Learn More