Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Object and scene image understanding method based on text-object-scene relations

An image understanding and object technology, applied in the field of image understanding of objects and scenes, to achieve the effect of saving manpower labeling overhead and improving accuracy

Active Publication Date: 2015-04-29
SHANGHAI JIAO TONG UNIV
View PDF6 Cites 11 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

However, the main disadvantage of this method is that it requires a one-to-one correspondence between objects and text instances to train the model, and it is necessary to manually mark the objects corresponding to the nouns during the prediction process.

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Object and scene image understanding method based on text-object-scene relations
  • Object and scene image understanding method based on text-object-scene relations
  • Object and scene image understanding method based on text-object-scene relations

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0041] The present invention will be described in detail below in conjunction with specific embodiments. The following examples will help those skilled in the art to further understand the present invention, but do not limit the present invention in any form. It should be noted that those skilled in the art can make several modifications and improvements without departing from the concept of the present invention. These all belong to the protection scope of the present invention.

[0042] The present invention fuses information of three domains: scene, object and text through a conditional random field. Features from different domains are related to each other through three relationships: scene to object, scene to text, and object to text. Text is associated with objects and scenes in two different ways. The relationship between text and objects is represented by the matching probability of nouns and objects, while the relationship between text and scenes is represented by ...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention relates to an object and scene image understanding method based on text-object-scene relations. The method comprises steps as follows: information of three domains are integrated through a conditional random field, and the three domains include a scene, an object and a text; features of different domains are correlated through three relations which include a scene and object relation, a scene and text relation and an object and text relation; the text is related with the object and the scene in two different manners. The relation of the text and the object is represented through matching probability of noun and the object, the relation of the text and the scene is represented through occurrence probability of the noun in description of different scenes; the matching probability of the text and the object is obtained through solution of a problem about bidirectional matching optimization constrained under the weak supervision condition; the method only requires simple annotated information, the accuracy is improved, network resources can be sufficiently utilized for image understanding, and huge manual annotation cost is saved.

Description

technical field [0001] The invention relates to a method in the technical field of computer vision for signal processing, in particular to an object and scene image understanding method based on text-object-scene relationship. Background technique [0002] In the era of big data, image data on the Internet has shown explosive growth, and an intelligent vision system that can automatically extract semantic information from images is urgently needed. After unremitting efforts in the field of computer vision, significant progress has been made in many key tasks, such as object detection, scene classification, and face recognition. However, as the ultimate goal in the field of computer vision, image understanding is still a very challenging problem. This problem becomes more difficult when faced with the massive image understanding on the Internet. People can no longer rely on accurately labeled image datasets to train models as in the past, because it is expensive to accurate...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(China)
IPC IPC(8): G06K9/62
CPCG06F18/241G06F18/2415
Inventor 熊红凯王博韬
Owner SHANGHAI JIAO TONG UNIV
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products