Unsupervised cross-modal hash retrieval method based on noisy label learning

A cross-modal, unsupervised technique, applied in the field of image hash retrieval, which can solve problems such as pseudo-label multi-noise

Active Publication Date: 2021-05-25
NANJING UNIV
View PDF11 Cites 4 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0005] Purpose of the invention: Aiming at the problem that the previous methods generally use the picture features extracted by the pre-trained network and the text features of the bag-of-words model to calculate the cosine similarity as a pseudo-label to directly fit, and the obtained pseudo-label contains a lot of noise. The present invention An unsupervised cross-modal hash retrieval method based on noisy label learning is provided. In the present invention, the idea of ​​learning with noisy labels is used ingeniously, and two sets of dual hash models are set, respectively using the "small error criterion "To select relatively clean and credible similarity labels in each small batch, feed them to the other party's neural network, and do training. Practice has proved that it can effectively improve the generalization performance of the model.

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Unsupervised cross-modal hash retrieval method based on noisy label learning
  • Unsupervised cross-modal hash retrieval method based on noisy label learning
  • Unsupervised cross-modal hash retrieval method based on noisy label learning

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0048] Below in conjunction with specific embodiment, further illustrate the present invention, should be understood that these embodiments are only used to illustrate the present invention and are not intended to limit the scope of the present invention, after having read the present invention, those skilled in the art will understand various equivalent forms of the present invention All modifications fall within the scope defined by the appended claims of the present application.

[0049] The following embodiments illustrate the method of the present invention by taking unsupervised image-text retrieval on social media networks as a specific example.

[0050] like figure 1As shown, the multimodal data acquisition steps are as follows: determine whether the text label is a specific description of the corresponding picture (step 100), and then organize all pictures and the text description corresponding to each picture into the form of "picture-text" pairs ( step 101).

[00...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention discloses an unsupervised cross-modal hash retrieval method based on noisy tag learning. The method comprises: a multi-modal data acquisition step, a dual multi-modal neural network construction step, a similarity pseudo label with noise extracted based on a pre-trained neural network, an unsupervised cross-modal Hash training step based on noise label learning and a final dual cross-modal Hash retrieval test step. According to the method, two dual Hash model groups are set and relatively clean pseudo labels are mutually fed to each other, so that the learning of the model is misguided by noise pseudo labels as little as possible, the final effect is that the model is converged to a better position, and the performance on a test data set is better than that of other unsupervised cross-modal Hash methods.

Description

technical field [0001] The invention relates to an unsupervised cross-modal hash retrieval method based on noisy label learning, which is used for image hash retrieval with text description. Background technique [0002] The general cross-modal hash retrieval task is implemented through a set of networks, including image networks and text networks. In the case of supervised labeling, each image / text is assigned multiple labels. If a picture and a text have at least one common label among their respective labels, they are called similar labels, otherwise they are called not similar. In the process of training this group of networks, a batch of "picture-text" pairs is sampled each time, and its feature representation is obtained after network processing. The similarity is calculated based on the feature representation, and the error function of the real similarity is calculated. The error function is propagated backwards, so that the network weights are updated in the direct...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
Patent Type & Authority Applications(China)
IPC IPC(8): G06F16/38G06F16/58G06F40/30G06K9/46G06K9/62G06N3/04G06N3/08
CPCG06F16/38G06F16/58G06F40/30G06N3/088G06N3/084G06V10/44G06N3/045G06F18/22
Inventor 詹德川杭诚王魏
Owner NANJING UNIV
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products