Cross-modality image-label relevance learning method facing social image

A correlation and cross-modal technology, applied in the field of cross-media correlation learning, can solve problems such as ignoring image visual information, calculating image and label correlation, ignoring image semantic information, etc., to achieve strong adaptability, high accuracy, The effect of high accuracy

Active Publication Date: 2015-09-09
FUDAN UNIV
View PDF3 Cites 45 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

These methods ignore the value of rich information in multimodal features to calculate the correlation between images and labels. Therefore, it is very necessary to find suitable algorithms to fully mine and construct multimodal feature representations [21,22] ,twenty three]
[0017] (2) Evaluation of multimodal correlation. Existing methods usually only consider single-modal information when calculating correlation, either ignoring the visual information of the image itself, or ignoring the semantic information contained in the image.
Instead of making full use of multimodal information to calculate the correlation between images and labels, rich multimodal information is used to calculate the correlation between images and labels

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Cross-modality image-label relevance learning method facing social image
  • Cross-modality image-label relevance learning method facing social image
  • Cross-modality image-label relevance learning method facing social image

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0077] The cross-modal relevance calculation method for social images of the present invention will be described in detail below in conjunction with the accompanying drawings.

[0078] (1) Collection data object

[0079] Collect data objects, obtain images and image annotation data, and organize image annotation data that do not appear frequently or are useless in the entire data set. Generally, the obtained data set contains a lot of noise data, so it should be properly processed and filtered before using these data for feature extraction. For images, the obtained images are all in a uniform JPG format, and no conversion is required. For text annotation of images, the resulting image annotations contain a lot of meaningless words, such as words plus numbers without any meaning. Some images have as many as dozens of annotations. In order for the image annotations to describe the main information of the image well, those useless and meaningless annotations should be discarded...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention belongs to the technical field of cross-media relevance learning, and particularly relates to a social image oriented cross-modality image-label relevance learning method. The invention comprises three algorithms: multi-modal feature fusion, bidirectional relevancy measuring and cross-modality relevancy fusion; a whole social image set is described by taking a hyperimage as a basic model, the image and a label are respectively mapped into hyperimage nodes for treatment, relevancy aiming at the image and the relevancy aiming at the label are obtained, and the two different relevancies are combined according to a cross-modality fusion method to obtain a better relevancy. Compared with the traditional method, the method is high in accuracy and high in adaptivity. The method has important significance in performing efficient social image retrieval by considering multi-modal semantic information based on large-scale social images with weak labels, retrieval relevancy can be improved, user experience is enhanced, and the method has application value in the field of cross-media information retrieval.

Description

technical field [0001] The invention belongs to the technical field of cross-media correlation learning, and in particular relates to a social image-oriented cross-modal image-label correlation learning method. technical background [0002] With the development of web 2.0 technology, especially the popularity of some social networking sites, such as Flickr. Ordinary users are more likely and willing to share image resources on the Internet. How to better deal with these massive social image data, effectively organize and manage its complex structure, and then promote the cross-media retrieval of these images has become an important research hotspot [1,2,3,4,5]. ,6]. Generally speaking, each social image will be marked as a series of tags, and these tags are provided by ordinary users, and these tags usually have the user's subjective views and tendencies [7]. However, due to the problem of semantic gap, there may be huge uncertainty between the visual content of the image...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
Patent Type & Authority Applications(China)
IPC IPC(8): G06F17/30
CPCG06F16/58
Inventor 张玥杰程勇刘志鑫金城张滨
Owner FUDAN UNIV
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products