Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Deep neural network model compression training method, device and apparatus and medium

A technology of deep neural network and training method, which is applied in the direction of biological neural network model, neural learning method, neural architecture, etc., and can solve problems such as limited application, high computational complexity, and reasoning speed that cannot meet real-time requirements

Pending Publication Date: 2020-08-04
SOUTH CHINA UNIV OF TECH
View PDF0 Cites 12 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0002] Driven by technologies such as large data sets and parallel computing, the neural network based on deep learning has received extensive attention from academia and industry, and has achieved relatively large breakthroughs in many fields, such as image classification, object detection, semantic segmentation, Face detection and face recognition, etc. However, the current algorithms based on deep neural network models still face many challenges
First of all, although the algorithm based on the deep neural network model has achieved relatively excellent recognition rates in multiple data sets, the current deep neural network model has a huge amount of parameters, which leads to huge storage and memory bandwidth requirements, resulting in too much resource usage. high
Secondly, the model based on the deep neural network has high computational complexity, which makes it difficult for the algorithm based on the deep neural network model to meet the real-time requirements in terms of inference speed, and is not suitable for devices with high real-time requirements.
The huge amount of parameters and calculations make it difficult to deploy algorithms based on deep neural network models to resource-constrained devices, such as mobile phones, wearable devices, and drones, etc., which greatly limits the use of algorithms based on deep neural network models in various fields. field application

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Deep neural network model compression training method, device and apparatus and medium
  • Deep neural network model compression training method, device and apparatus and medium
  • Deep neural network model compression training method, device and apparatus and medium

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0059] The technical solutions in the embodiments of the present application will be clearly and completely described below in conjunction with the accompanying drawings in the embodiments of the present application. Obviously, the described embodiments are only some of the embodiments of the present application, not all of them. Based on the embodiments in this application, all other embodiments obtained by persons of ordinary skill in the art without making creative efforts belong to the scope of protection of this application.

[0060] see figure 1 As shown, the embodiment of the present application discloses a deep neural network model compression training method, the method comprising:

[0061] Step S11: Obtain the target training data set.

[0062] In a specific implementation process, a target training data set needs to be obtained first, wherein the target training data set includes but not limited to a target image training set. When the target training data set is ...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses a deep neural network model compression training method, device and apparatus, and a medium. The method comprises the steps of obtaining a target training data set; inputting the target training data set into a pre-constructed first deep neural network model and a pre-constructed second deep neural network model to obtain a first output and a second output; constructing a target loss function according to the first output and the second output; updating a first full-precision weight parameter of the first deep neural network model and a second full-precision weight parameter of the second deep neural network model by using the target loss function; and updating a quantization weight parameter of the second deep neural network model by utilizing the second full-precision weight parameter, and taking the second deep neural network model as a trained compression neural network model when the target loss function meets a preset requirement. Therefore, the size of the model can be reduced, the storage and memory bandwidth requirements are reduced, and the calculation cost is reduced.

Description

technical field [0001] The present application relates to the technical field of machine learning, in particular to a deep neural network model compression training method, device, equipment, and medium. Background technique [0002] Driven by technologies such as large data sets and parallel computing, the neural network based on deep learning has received extensive attention from academia and industry, and has achieved relatively large breakthroughs in many fields, such as image classification, object detection, semantic segmentation, Face detection and face recognition, etc. However, current algorithms based on deep neural network models still face many challenges. First of all, although the algorithm based on the deep neural network model has achieved relatively excellent recognition rates in multiple data sets, the current deep neural network model has a huge amount of parameters, which leads to huge storage and memory bandwidth requirements, resulting in too much resou...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(China)
IPC IPC(8): G06N3/08G06N3/04
CPCG06N3/082G06N3/047G06N3/045
Inventor 吴庆耀刘璟谭明奎
Owner SOUTH CHINA UNIV OF TECH
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products