Cross-modal retrieval method based on adversarial learning and asymmetric hashing

A cross-modal and asymmetric technology, applied in the field of cross-modal retrieval, which can solve the problem that the hash code is not optimal, and the modal data cannot be fully extracted.

Active Publication Date: 2019-09-10
INST OF INFORMATION ENG CAS
View PDF11 Cites 28 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

However, the existing cross-modal methods based on deep learning cannot fully extract the characteristics of each moda

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Cross-modal retrieval method based on adversarial learning and asymmetric hashing
  • Cross-modal retrieval method based on adversarial learning and asymmetric hashing
  • Cross-modal retrieval method based on adversarial learning and asymmetric hashing

Examples

Experimental program
Comparison scheme
Effect test

Example Embodiment

[0051] In order to better express the cross-modal retrieval method based on adversarial learning and asymmetric hashing proposed in the present invention, in conjunction with the drawings and specific implementations, a 224×224 pixel picture and its corresponding text description are taken as an example. , To further explain the present invention.

[0052] figure 1 It is the overall flow chart of the present invention, including five parts: data preprocessing, initialization model framework, model training, hash code generation of each modal data, and retrieval stage.

[0053] Step 1. Data preprocessing. Divide the cross-modal data set into two parts: training set and test set. Each data instance contains picture-text pairs and their corresponding labels;

[0054] Step 2. Initialize the model framework. figure 2 It is a model framework designed in the present invention. The framework includes a cross-modal feature extraction module, an attention module guided by a counter-network,...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention discloses a cross-modal retrieval method based on adversarial learning and asymmetric hashing. The cross-modal retrieval method comprises the following steps: 1) selecting a cross-modaldata set and dividing the cross-modal data set into a training set and a test set; training a model, wherein the model comprises a loss module, a feature extraction unit and a Hash code learning unit,wherein the hash code learning unit comprises a hash code learning network and a shared multi-label binary matrix, and the loss module comprises the confrontation loss of the feature extraction unit,the ternary margin loss of the hash code learning unit, the cosine quantization loss and the asymmetric hash loss; 2) generating a binary hash code of each modal data in the cross-modal data set by using the trained optimal model; and 3) for a given query data, firstly generating a binary hash code, and then calculating a Hamming distance between the binary hash code of the query data and the binary hash code of the modal data in the cross-modal data set, which is different from the modal of the query data, so as to obtain a cross-modal data instance meeting the condition.

Description

technical field [0001] The invention relates to a cross-modal retrieval method based on adversarial learning and asymmetric hashing, and belongs to the technical field of computer software. Background technique [0002] With the massive increase of multimedia data in online social media and search engines, there is an increasing demand for mutual retrieval between different modal data (such as video, image, text, etc.). For example, when a user inputs a text query, it needs to obtain picture information or audio and video information related to the query text. In order to solve the retrieval problem in large-scale cross-modal data, an efficient and fast method is the hash method. The cross-modal hash method mainly has two stages, one is the extraction of the characteristics of each modal data, and the other is It is the similarity mining between various modal data. Currently existing cross-modal hashing methods can be roughly divided into two types: traditional cross-modal...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
IPC IPC(8): G06F16/31G06F16/335G06F16/58G06F16/583G06F16/51G06N3/04G06N3/08
CPCG06F16/325G06F16/335G06F16/5846G06F16/5866G06F16/51G06N3/08G06N3/045
Inventor 古文李波古晓艳熊智谷井子王伟平
Owner INST OF INFORMATION ENG CAS
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products