Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Compression acceleration method of deep convolutional neural network for target detection

A deep convolution and neural network technology, applied in the field of deep learning and artificial intelligence, can solve problems such as limiting the application of convolutional neural network, and achieve the effect of saving calculation amount

Pending Publication Date: 2020-05-05
INST OF SEMICONDUCTORS - CHINESE ACAD OF SCI
View PDF0 Cites 6 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0003] However, because the convolutional neural network is accompanied by a large amount of storage and calculation, for example, the classic VGG16 network requires about 520MB of storage and 15.3 billion multiplication and addition operations. Even some existing lightweight networks still require dozens of Megabytes of storage and millions of multiplication and addition operations, such a huge amount of storage and calculations limit the application of convolutional neural networks, especially in mobile devices and embedded devices

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Compression acceleration method of deep convolutional neural network for target detection
  • Compression acceleration method of deep convolutional neural network for target detection
  • Compression acceleration method of deep convolutional neural network for target detection

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0036] In order to make the object, technical solution and advantages of the present invention clearer, the present invention will be further described in detail below in conjunction with specific embodiments and with reference to the accompanying drawings.

[0037] See figure 1 , the present invention comprises the following four steps:

[0038] Step 1: Construct and train a deep convolutional neural network for target detection;

[0039] Step 2: Quantify and test all weight values ​​in the deep convolutional neural network and the activation values ​​of each layer except the last layer after the activation function. The quantization step size is from small to large, and test the deep convolutional neural network Detect performance loss, and select the largest quantization step within the set loss range;

[0040] Step 3: Using the above-mentioned maximum quantization step size, according to the number of compression bits required by the network, determine the truncated rang...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses a compression acceleration method of a deep convolutional neural network for target detection. The method comprises the following steps: constructing and training a deep convolutional neural network for target detection; carrying out quantitative test on all weight values in the deep convolutional neural network and activation values of all layers except the last layer after passing through an activation function, testing the detection performance loss condition of the network from small to large in quantitative step size, and selecting the maximum quantitative step size in a set loss range; determining a truncation range of a weight value and an activation value in the neural network by utilizing the quantization step length, limiting the neural network and training the network; and truncating and quantifying the deep convolutional neural network, and compiling a forward code. According to the method, the quantization technology is adopted to reduce the networkstorage capacity, 32-bit floating-point number operation in the network is converted into 8-bit integer operation, and meanwhile, the sparsity of the network is utilized to convert the layer meetingthe sparsity condition in the network into sparse matrix operation, so that the purpose of compressing and accelerating the deep convolutional neural network is achieved.

Description

technical field [0001] The invention relates to the fields of deep learning and artificial intelligence, in particular to a compression acceleration method of a deep convolutional neural network for target detection. Background technique [0002] Object detection has always been an important research direction in the field of computer vision due to its broad application prospects in information retrieval, autonomous driving, robot navigation, and augmented reality, and has received extensive research and attention from academia and industry. The traditional object detection system mainly uses some artificially designed features, such as Haar features and HoG features, etc., and uses classifiers such as support vector machines to classify images with sliding windows to achieve the effect of object detection. In recent years, with the rise of deep learning, convolutional neural networks have brought extremely effective solutions to target detection. The results obtained by met...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(China)
IPC IPC(8): G06N3/04G06N3/063G06N3/08
CPCG06N3/082G06N3/063G06N3/048G06N3/045
Inventor 李志远余成宇吴绮金敏鲁华祥陈艾东郭祉薇
Owner INST OF SEMICONDUCTORS - CHINESE ACAD OF SCI
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products