Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

An Image Classification Method Based on Dimension Transformation of Observation Matrix

A technology of observation matrix and classification method, applied in instrument, calculation, character and pattern recognition, etc., can solve the problems of unfavorable storage and transmission, increase of calculation amount, low model parameter adjustment training and use efficiency, etc., to improve the efficiency of the model. Effect

Active Publication Date: 2020-12-25
ZHEJIANG UNIV
View PDF5 Cites 0 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

After full sampling, the data volume of the digital signal is relatively large, which is not conducive to storage and transmission on the one hand, and on the other hand, the digital signal itself has a lot of redundancy
At present, in the field of image classification, with the increase of input data and model complexity, the amount of calculation is increasing day by day, and the efficiency of model tuning training and use is not high.

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • An Image Classification Method Based on Dimension Transformation of Observation Matrix
  • An Image Classification Method Based on Dimension Transformation of Observation Matrix
  • An Image Classification Method Based on Dimension Transformation of Observation Matrix

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0045] The present invention will be further described below in conjunction with the accompanying drawings and embodiments.

[0046] Such as figure 1 As shown, this embodiment provides an image classification method based on the transformation dimension of the observation matrix, including the following steps:

[0047] (1) Use perceptual compression to sparsely encode images to obtain a data set composed of low-dimensional images. Divide the data set containing labels into training set and test set, and the division ratio is 8:2.

[0048] Methods for sparsely coding images using perceptual compression include sparse representation of images, image compression sampling, and image reconstruction.

[0049] (1-1) Image sparse representation is:

[0050] Express the original signal x on a set of sparse basis Ψ:

[0051] x=Ψs

[0052] Among them, x is the original signal, its size is N×1, Ψ is a set of sparse basis, and s is the sparse coefficient.

[0053] s is an N×1 column v...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses an image classification method based on transforming dimensions of an observation matrix, which includes: using perceptual compression to sparsely encode images to obtain a data set composed of low-dimensional images, and dividing the data set containing labels into a training set and a test set Constructing an image classification network comprising an input layer, a hidden layer and an output layer, the hidden layer is a perceptron unit; the image classification network is at least two, respectively comprising different node number perceptron units; the training set is used as Input, train under the supervision of the label, and obtain the corresponding neural network image classification model after training; use the test set to verify the accuracy of the neural network image classification model for image classification, and select the one with the highest accuracy as the final neural network image classification Model; input the image to be tested, and output the predicted probability of the image classification result. The image classification method provided by the present invention can greatly improve model efficiency without reducing image classification accuracy.

Description

technical field [0001] The invention belongs to the technical field of image classification, in particular to an image classification method based on transformation dimension of observation matrix. Background technique [0002] Image classification is an image processing method that distinguishes different types of objects according to the different characteristics reflected in the image information. It uses computer to carry out quantitative analysis on images, and classifies each pixel or area in the image or image into one of several categories to replace human visual interpretation. [0003] Classification method of image space—use image grayscale, color, texture, shape, position and other underlying features to classify images; for example: [1] use gray histogram features to classify images; [2] use texture features to classify images Image classification; [3] uses texture, edge and color histogram mixed features to classify images; [1], [2], [3] all use SVM as a class...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Patents(China)
IPC IPC(8): G06K9/62
CPCG06F18/213G06F18/24G06F18/214
Inventor 叶心汝王勇
Owner ZHEJIANG UNIV
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products