Method and device for extracting hash code from image and image retrieval method and device

An image-in-redundant hash code technology, applied in still image data retrieval, metadata still image retrieval, image coding, etc., can solve the problems of image retrieval accuracy decline, hidden layer unit collapse, insufficient coding space utilization, etc. , to achieve the effect of improving the accuracy and reducing the redundancy of coding spatial information

Active Publication Date: 2021-07-13
BEIJING QIHOO TECH CO LTD
View PDF4 Cites 0 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0004] The variational pruning of the VAE framework will cause some hidden layer units to collapse when they are not effectively extracted in the early stage of model training, so that the framework has obvious inherent deficiencies, for example, (1) There are many coding spaces Redundant dimension (i.e., redundant data without information); (2) The framework underutilizes the latent code of the coding space; etc.
Especially when the decoder structure is complex, these shortcomings are more obvious
This will lead to: the inability to accurately extract the image hash code, resulting in a decrease in the accuracy of image retrieval and other related application problems

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Method and device for extracting hash code from image and image retrieval method and device
  • Method and device for extracting hash code from image and image retrieval method and device
  • Method and device for extracting hash code from image and image retrieval method and device

Examples

Experimental program
Comparison scheme
Effect test

Embodiment 1

[0152] Embodiment 1, MNIST data and image reconstruction

[0153] Network parameter settings: the number of encoder and decoder layers M is set to 1, δ and η are 0.01, and the prior parameters ρ j is 0.5, the threshold parameter is set to 0.05, the encoder input data and decoder output data dimensions are 28*28=784, the hash code and the hidden layer data dimensions of the encoder and decoder are 64;

[0154] The training data set is the training set of MIST; the batch size of data samples used in each step of training is set to 32; the image reconstruction error of the model is evaluated at different numbers of training rounds, and the evaluation uses the MIST test set, which is extracted by the encoder. The code and decoder generate reconstructed data, and calculate the error between input and reconstructed data. The calculation method is as follows:

[0155]

[0156] Where N is the number of evaluation samples, D is the dimensionality of each sample data, x is the input...

Embodiment 2

[0158] Embodiment 2, CIFAR-10 image retrieval

[0159] Network parameter settings: the number of encoder and decoder layers M is set to 4, δ is 0.01, η is 0.01, the prior parameter ρj is 0.5, the threshold parameter is set to 0.05, the dimension of encoder input data and decoder output data is 512, The data dimensions of the hash code and the hidden layers of the encoder and decoder are 32, 64 and 128;

[0160] 100 sample data are randomly selected from each of the 10 types of data in the CIFAR-10 dataset, a total of 1000 sample data are used as the retrieval input during testing, and the rest of the data are training samples and image databases. For each step of training, the batch size of data samples is set to 32, and the number of training rounds is 200.

[0161] Use the mAP index to evaluate the image retrieval ability. The mAP results of the three models with hash code dimensions of 32, 64 and 128 are shown in Table 1. Table 1 shows the mAP (%) test results of SGH and R...

Embodiment 3

[0165] Embodiment 3, Caltech-256 image retrieval

[0166] Network parameter settings: the number of encoder and decoder layers M is set to 4, δ is 0.01, η is 0.01, and the prior parameter ρ j is 0.5, the threshold parameter is set to 0.05, the encoder input data and decoder output data dimensions are 512, the hash code and the hidden layer data dimensions of the encoder and decoder are 32, 64 and 128;

[0167] 1000 sample data are randomly selected from the Caltech-256 dataset as the retrieval input during testing, and the rest of the data are training samples and image libraries. For each step of training, the batch size of data samples is set to 32, and the number of training rounds is 200.

[0168]Use the mAP index to evaluate the image retrieval ability. The mAP results of the three models with hash code dimensions of 32, 64 and 128 are shown in Table 2. Table 2 shows the mAP (%) test results of SGH and R-SGH in the Caltech-256 data set :

[0169] Table 2 SGH and R-SGH ...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention discloses a method and device for extracting a hash code from an image, and an image retrieval method and device. The method includes: constructing a hash code extraction model, the model includes an encoder and a decoder; wherein, the encoder consists of multiple The layer deep neural network DNN is composed, and the hash code is extracted from the image data and output to the decoder. The decoder is composed of multiple layers of DNN, and the input hash code is converted into an image; in the encoder, the hidden layer coding redundancy is measured. Regularize the hidden layer coding to reduce the redundancy of coding space information, and obtain the anti-redundancy hash code depth extraction model; train the anti-redundancy hash code depth extraction model to determine the parameters in the model; use the trained anti-redundancy hash code depth extraction model Encoder in Redundant Hash Code Deep Extraction Model to extract hash codes from images. This method can effectively reduce the redundancy of coding space information, effectively utilize all dimensions, extract image hash codes with high precision, and effectively improve the accuracy of related application fields such as image retrieval.

Description

technical field [0001] The present invention relates to the technical field of artificial intelligence, in particular to a method and device for extracting a hash code from an image, an image retrieval method and device, electronic equipment and a computer-readable storage medium. Background technique [0002] LTH (learning to hash) is an image compression method that is very effective in image retrieval applications. This framework extracts binary hash codes from images, calculates the similarity between the input image and the image hash codes in the image library, and performs retrieval. The LTH framework can greatly reduce storage space and improve retrieval efficiency. [0003] The hash code extraction of the image in LTH is very critical, and it is generally implemented by an encoder. An autoencoder is an unsupervised neural network method consisting of an encoder and a decoder that generates images from a random code. VAE (Variational Autoencoder, Variational Autoen...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
Patent Type & Authority Patents(China)
IPC IPC(8): G06F16/58G06F16/53G06T9/00
CPCG06T9/002
Inventor 王浩杜长营庞旭林张晨杨康
Owner BEIJING QIHOO TECH CO LTD
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products