Specific modal semantic space modeling-based cross-modal similarity learning method

A technique for modal semantics, learning methods

Active Publication Date: 2018-01-09
PEKING UNIV
View PDF4 Cites 81 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

Therefore, treating different modality data equally to mine the potential fine-grained alignment content and construct a unified space will lose the unique and useful information within the modality, instead of making full use of the rich intrinsic information provided by each modality.

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Specific modal semantic space modeling-based cross-modal similarity learning method
  • Specific modal semantic space modeling-based cross-modal similarity learning method
  • Specific modal semantic space modeling-based cross-modal similarity learning method

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0021] The present invention will be further described in detail below in conjunction with the accompanying drawings and specific embodiments.

[0022] A cross-modal similarity learning method based on specific modality semantic space modeling of the present invention, its process is as follows figure 1 shown, including the following steps:

[0023] (1) Establish a cross-modal database, which contains data of multiple modality types, and divide the data in the database into training set, test set and validation set.

[0024] In this embodiment, the cross-modal database may contain multiple modality types, including images and texts.

[0025] Let D denote the cross-modal data set, D={D (i) ,D (t)},in

[0026] For media type r, where r=i, t (i means image, t means text), define n (r) for the number of data. Each data in the training set has one and only one semantic category.

[0027] definition is the feature vector of the pth data in the media type r, and its repres...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention relates to a specific modal semantic space modeling-based cross-modal similarity learning method. The method comprises the following steps of: 1, establishing a cross-modal database which comprises data of multiple modal types, and dividing data in the database into a training set, a test set and a validation set; 2, aiming at each modal type in the cross-modal database, constructinga semantic space aiming at the specific modal, and projecting data of the other modal types to the semantic space so as to obtain a cross-modal similarity aiming at the specific modal; 3, fusing cross-modal similarities, aiming at specific modals, obtained from different modal semantic spaces to obtain a final cross-modal similarity; and 4, calculating a similarity between a query example and a query target by taking any modal type in the test set as a query modal and taking another modal type as a target modal, and obtaining a related result list of data of the target modal according to thesimilarity. According to the method, the correctness of cross-modal retrieval can be improved.

Description

technical field [0001] The invention relates to the field of multimedia retrieval, in particular to a cross-modal similarity learning method based on specific modality semantic space modeling. Background technique [0002] Nowadays, multimodal data including images, videos, texts and audios widely exist on the Internet, and these multimodal data are the basis for helping artificial intelligence to recognize the real world. Some research work has been trying to break the heterogeneous gap between different modal data, and cross-modal retrieval, as one of the hot research issues, can realize information retrieval across different modal data, and has a wide range of practical application requirements. Such as search engines and digital libraries. Traditional single-modal retrieval, such as image retrieval, video retrieval, etc., is limited to a single modality, and can only return retrieval results of the same modality type as the query. The difference is that cross-modal ret...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
Patent Type & Authority Applications(China)
IPC IPC(8): G06F17/30G06N3/08
Inventor 彭宇新綦金玮
Owner PEKING UNIV
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products