Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Object semantics and deep appearance feature fusion-based scene identification method

A feature fusion and scene recognition technology, applied in the field of image processing, can solve the problems of small differences between classes, the decline of recognition effect, and large differences in indoor scenes, so as to improve the recognition rate and robustness, and improve the recognition accuracy rate. Effect

Active Publication Date: 2018-07-27
NANJING UNIV OF POSTS & TELECOMM
View PDF3 Cites 19 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0005] However, when the above method is applied to indoor scene recognition, the recognition effect declines. The reason is that indoor scenes mainly have intra-class differences, large inter-class differences, etc.

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Object semantics and deep appearance feature fusion-based scene identification method
  • Object semantics and deep appearance feature fusion-based scene identification method

Examples

Experimental program
Comparison scheme
Effect test

Embodiment 1

[0031] see figure 1 , a scene recognition method based on the fusion of target semantics and deep appearance features provided by this embodiment, the specific steps are:

[0032] Obtain the scene image to be recognized;

[0033] Extract the target semantic information of the scene image, and generate the target semantic features that maintain the spatial layout information;

[0034] Extract the appearance context information of the scene image to generate appearance context features;

[0035] Extract the appearance global information of the scene image, and generate the appearance global feature;

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses an object semantics and deep appearance feature fusion-based scene identification method. The method comprises the specific steps of obtaining a to-be-identified scene image; extracting object semantics information of the scene image to generate an object semantics feature for keeping spatial layout information; extracting appearance context information of the scene image to generate an appearance context feature; extracting global appearance information of the scene image to generate a global appearance feature; and according to the object semantics feature, the appearance context feature and the global appearance feature, obtaining an identification result of the scene image. By adopting a multi-class object detector algorithm, a key object, a class and layout information are accurately obtained; the object semantics feature of the indoor scene image is obtained through an SFV model; and convolutional layers and LSTM layers form an end-to-end trainable hybridDNN structure, so that the context information of the scene image can be effectively extracted. The method fuses the object semantics information, the global appearance information and the appearancecontext feature, so that the identification rate and robustness of the identification method are improved.

Description

technical field [0001] The invention belongs to the technical field of image processing, and in particular relates to a scene recognition method based on fusion of target semantics and deep appearance features. Background technique [0002] Scene recognition is one of the important topics of computer vision, and it is widely used in many fields, mainly including image information retrieval of large databases, mobile positioning of robots and interaction with the environment, event detection in the field of security monitoring, etc. [0003] Since 2006, deep learning theory has become a research hotspot in the field of machine learning and artificial intelligence. Deep learning establishes a deep network structure to simulate the cognitive mechanism of the human brain, and extracts input data layer by layer through a deep multi-layer network to form The mapping relationship from low-level signals to high-level semantics. Deep learning architecture has achieved great success ...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
IPC IPC(8): G06K9/00G06K9/46G06K9/62G06N3/04
CPCG06V20/36G06V10/424G06V2201/07G06N3/045G06F18/2135G06F18/2411G06F18/253G06F18/214
Inventor 孙宁李文丽李晓飞
Owner NANJING UNIV OF POSTS & TELECOMM
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products