Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

A feature image recognition method based on a multi-attention space pyramid

A space pyramid and feature image technology, applied in the field of improved deep convolutional network structure, can solve the problems of difficult fine-grained feature extraction and low recognition accuracy of images, and achieve enhanced feature extraction capabilities, improved accuracy, and improved The effect of accuracy

Inactive Publication Date: 2019-06-18
TAIYUAN UNIV OF TECH
View PDF7 Cites 37 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0005] In order to improve the performance of the network, so that it has better robustness and higher recognition accuracy in complex environments, it solves the problem that the accuracy of image recognition with low pixels and in complex environments is not high, and the image It is difficult to extract the fine-grained features of the environment, and a feature image recognition method based on the multi-attention space pyramid is proposed to strengthen the feature extraction ability of the convolutional neural network and realize fine image recognition in complex environments.

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • A feature image recognition method based on a multi-attention space pyramid
  • A feature image recognition method based on a multi-attention space pyramid
  • A feature image recognition method based on a multi-attention space pyramid

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0014] figure 1 As shown, in the image input layer, because the model adds spatial pyramid pooling, feature maps of any size can be converted into feature vectors of fixed size, and there is no requirement for the size of the input image, and images of any size can be input. Feature extraction stage:

[0015] Constructing a feature extraction network based on a multi-attention space pyramid is as follows: Based on the Inceptionv3 network, a feature extraction network is proposed. The feature extraction network has a main network and three branch networks, and each branch network shares the convolutional layer of the CNN model. Each branch has the same inception module as the main network, such as figure 1 Among them, the CNN structure contains five convolutional layers, two average pooling layers, and the Relud activation function and BN (standardized) operation are added after each convolution, specifically:

[0016] The convolution kernel size is 3×3, the depth is 32, the ...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses a feature image recognition method based on a multi-attention space pyramid, and belongs to the technical field of network mode identification. The method is characterized in that a shallow network deepening method is combined; the feature extraction capability of the convolutional neural network is enhanced; More abundant feature representations are obtained by fusing multi-scale attention features through an attention module, and a spatial pyramid pooling operation is added at the end of each branch network to perform multi-scale feature extraction, so that feature maps of any size are converted into feature vectors of fixed sizes to be sent to a full connection layer. The network provided by the invention can input images with any size without zooming or cutting,better retains the feature information of the images, and has better robustness and accuracy in the aspects of image fine recognition and low-pixel image classification recognition in a complex environment.

Description

technical field [0001] The invention belongs to the technical field of network pattern recognition, a multi-attention space pyramid feature extraction network structure, combined with a method of deepening the shallow network, strengthens the feature extraction ability of the convolutional neural network, and is an improved deep convolutional network structure . Background technique [0002] With the continuous development of mobile Internet and pattern recognition technology, the transmission of information has become faster, and the rapid increase in the amount of information has made the communication medium gradually shift from text to pictures or videos. Compared with the time-consuming and labor-intensive traditional image processing, deep learning has become a research hotspot in the fields of image recognition and artificial intelligence with its powerful data processing capabilities and high accuracy. [0003] Among them, the convolutional neural network is a data-...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(China)
IPC IPC(8): G06K9/46G06N3/04
Inventor 段迅达王楷元其他发明人请求不公开姓名
Owner TAIYUAN UNIV OF TECH
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products