Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Scene map generation method

A technology of scene graph and image area, which is applied in the direction of neural learning method, biological neural network model, editing/combining graphics or text, etc. It can solve the problem of not using the relevant information of the scene graph, being unable to understand the main content of the scene image, and having no prominent display Image visual relationship and other issues

Active Publication Date: 2020-07-28
HARBIN ENG UNIV
View PDF10 Cites 17 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

The existing traditional scene graph generation methods only perform object detection and relationship prediction on part of the content of the entire scene image, but cannot understand the main content of the scene image in a targeted manner.
At this stage, convolutional neural networks and recurrent neural network structures are used to generate image descriptions. The existing image description methods are only based on the target objects in the scene image and combined with natural language processing to generate an overall language description of the image content, and do not take advantage of the scene The relevant information of the image does not highlight the visual relationship between the objects to be expressed in the image

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Scene map generation method
  • Scene map generation method
  • Scene map generation method

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0045] The specific embodiments of the present invention will be further described below in conjunction with the accompanying drawings.

[0046] refer to figure 1 Specifically explaining this embodiment, the multi-level semantic task-based scene graph generation method described in this embodiment mainly includes Faster R-CNN object feature extraction, information transfer graph, feature information iterative update, image region description, and scene graph generation.

[0047] 1. For the three different levels of semantic vision tasks in scene understanding, object detection, visual relationship detection, and image region description, three different sets of proposals are produced:

[0048] Object region proposal: use the Faster R-CNN network to detect objects on the input image, and extract a set of candidate regions from the input image B={b 1 , b 2 ,...,b n}. For each region, the model not only extracts the bounding box b i Represents the position of the object, and...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses a scene graph generation method, which is characterized in that three different levels of semantic tasks of object detection, visual relationship detection and image region description are mutually connected, and the visual tasks of different semantic levels of scene understanding are jointly solved in an end-to-end mode. Firstly, an object, a visual relationship and imageregion description are aligned with a feature information transfer graph according to spatial features and semantic connection of the object, the visual relationship and the image region description,and then feature information is transferred to three different levels of semantic tasks through the feature information transfer graph so as to achieve simultaneous iterative updating of semantic features. According to the method, object detection and visual relation detection are realized by utilizing semantic feature connection of different levels of scene images so as to generate scene images corresponding to the scene images; and the main region of the scene image is subjected to image description by using a natural language, Meanwhile, the image region description is used as a supervisionmethod for scene image generation so as to improve the accuracy of scene image generation.

Description

technical field [0001] The invention relates to a method for generating a scene graph, in particular to a method for generating a scene graph based on multi-level semantic tasks, belonging to the fields of object detection, visual relationship detection and image region description. Background technique [0002] Scene understanding is one of the hot issues in computer vision research. Visual scene understanding includes multiple semantic tasks at different levels: object detection and recognition, prediction of the visual relationship between detected objects, and the use of natural language to express the content of the scene image. description etc. Due to the complexity and diversity of object categories and their visual relationships, visual scene understanding is still a difficult problem. A good visual scene understanding system can not only identify the differences in the content represented by different images, but also focus on or represent the differences. . The k...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(China)
IPC IPC(8): G06T11/60G06N3/04G06N3/08
CPCG06T11/60G06N3/08G06N3/044G06N3/045
Inventor 莫宏伟田朋姜来浩许贵亮杨帆
Owner HARBIN ENG UNIV
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products