Big data cross-modal retrieval method and system based on deep integration Hash

A big data and hashing technology, applied in the field of big data cross-modal retrieval methods and systems, can solve problems that cannot alleviate the heterogeneity of different modalities, cannot generate high-quality, compact hash codes, and cannot effectively capture images. problems such as spatial dependencies and temporal dynamics of sentences, to achieve the effect of reducing heterogeneity and improving accuracy

Inactive Publication Date: 2018-04-03
TSINGHUA UNIV
View PDF3 Cites 47 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0004] However, these techniques cannot effectively capture the spatial dependence of images and the temporal dynamics of sentences, so they cannot learn powerful feature representations and cross-modal embeddings, and cannot alleviate the heterogeneity of different modalities, so they cannot generate images for cross-modal High-quality, compact hash coding for modal retrieval, which does not perform well in cross-modal retrieval

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Big data cross-modal retrieval method and system based on deep integration Hash
  • Big data cross-modal retrieval method and system based on deep integration Hash
  • Big data cross-modal retrieval method and system based on deep integration Hash

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0047] The specific implementation manners of the present invention will be further described in detail below in conjunction with the accompanying drawings and embodiments. The following examples are used to illustrate the present invention, but are not intended to limit the scope of the present invention.

[0048] In the present invention, the terms "first" and "second" are only used to describe the difference, and should not be understood as indicating or implying relative importance. The term "plurality" means two or more, unless otherwise clearly defined.

[0049] figure 1 It is a flow chart of a deep fusion hash-based cross-modal retrieval method for big data provided by an embodiment of the present invention, as shown in figure 1 As shown, the method includes:

[0050] S1. Identify the data type of the data to be retrieved. If the data type of the data to be retrieved is an image, input the image to be retrieved into the trained image hash network, and obtain the imag...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention provides a big data cross-modal retrieval method based on deep integration Hash. The method comprises the following steps that: identifying the data type of data to be retrieved, if thedata type of the data to be retrieved is an image, inputting an image to be retrieved into a trained image Hash network, and obtaining an image binary system code corresponding to the image to be retrieved; calculating a first Hamming distance between the image binary system code corresponding to the image to be retrieved and the binary system code corresponding to each statement in a search library; and selecting a preset quantity of statements with a minimum Hamming distance from the search library as the data type of the data to be retrieved as an image retrieval result. By use of the big data cross-modal retrieval method and system based on the deep integration Hash, an internal cross corresponding relationship between visual data and natural language is captured, so that the compact Hash code of the image and the statement is generated in an end-to-end deep learning framework, and cross-modal retrieval accuracy is improved.

Description

technical field [0001] The present invention relates to the technical field of computer data management, and more specifically, to a large data cross-modal retrieval method and system based on deep fusion hashing. Background technique [0002] With the rapid development of information technology, high-dimensional multimedia data in search engines and social networks continues to increase. How to perform efficient cross-modal approximate nearest neighbor search in these massive high-dimensional big data has become an important and urgent problem. Cross-modal retrieval uses data from one modality as a query condition to return relevant results from another modality, such as retrieving images with sentences or retrieving sentences with images. A powerful solution for efficient cross-modal approximate nearest neighbor search is the hashing method (Hashing), which compresses high-dimensional data into compact binary codes, thus greatly improving its storage efficiency and retriev...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
Patent Type & Authority Applications(China)
IPC IPC(8): G06F17/30
CPCG06F16/325G06F16/3334G06F16/355G06F16/36G06F16/5838
Inventor 王建民龙明盛曹越刘斌
Owner TSINGHUA UNIV
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products