Multi-modal image target detection method based on image fusion

A multi-modal image, image fusion technology, applied in the fields of deep learning, computer vision and image fusion, can solve problems such as lack of structural features

Active Publication Date: 2019-10-11
TIANJIN UNIV
View PDF13 Cites 35 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

In this method, the image fusion network is used as the pre-order step of the target detection model, and a general target detection method suitable for infrared images is proposed. The fused image with infrared and visible light image features is detected by the target detection model based on deep neural network, so as to overcome the problem of lack of structural features of a single infrared sensor, which is of great significance to the improvement of detection results and practical engineering applications

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Multi-modal image target detection method based on image fusion
  • Multi-modal image target detection method based on image fusion
  • Multi-modal image target detection method based on image fusion

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0029] In order to make the technical solution of the present invention clearer, the specific embodiments of the present invention will be further described below in conjunction with the accompanying drawings. The specific implementation plan flow chart is as follows figure 1 shown.

[0030] The goal of the fusion network work of this scheme is to learn a mapping function based on the structure of the generative confrontation network. This function generates a fusion image based on two input images given by multiple unlabeled sets, which are the visible light input image v and the infrared input image u. The network is not limited to image domain translation between two images, but can be used on unlabeled image sets for fusion tasks.

[0031] The fused image can not only preserve the high contrast between the target and the background in the infrared image, but also retain more texture details than the source image. Similar to the sharpened infrared image, the fused image h...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention relates to a multi-modal image target detection method based on image fusion, and the method comprises the steps: 1), making a multi-modal image data set through an infrared image and avisible light image which are collected in advance; 2) taking the preprocessed paired images as the input of a generation model G in the fusion model; generating a model G based on U-Net full convolutional network. A convolutional neural network based on a residual network is used as a generative network model structure and comprises a contraction process and an expansion process, a contraction path comprises a plurality of convolution plus ReLU active layer plus maximum pooling (Max Pooling) structures, the number of feature channels of each step of down-sampling is doubled, and a generated fusion image is output; the fused image is input into a discrimination network model in a fusion model; according to the change of a loss function in a training process, a learning rate training indexis adjusted according to the number of iterations, and through training, based on a self-owned multi-modal image data set, an image fusion model which retains infrared image thermal radiation characteristics and visible light image structural texture characteristics at the same time can be obtained.

Description

technical field [0001] The invention belongs to the fields of deep learning, computer vision and image fusion, and relates to a deep neural network-based infrared-visible multimodal image fusion model and a target detection method for a target detection model. Background technique [0002] In the natural environment, objects will radiate electromagnetic waves of different frequencies that cannot be seen by the human eye, which is called thermal radiation [1]. The infrared images taken by the infrared sensor can record the thermal radiation of different objects. Compared with Visible Image (VI) images, infrared (Infrared Image, IR) images have the following characteristics: they can reduce the influence of external environments such as sunlight and smoke [1]; they are sensitive to objects and areas with obvious infrared thermal characteristics. At present, target detection tasks in infrared images are widely used, including important applications in military, electric power,...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
Patent Type & Authority Applications(China)
IPC IPC(8): G06T5/50G06T7/00G06N3/04
CPCG06T5/50G06T7/0002G06T2207/10048G06T2207/10024G06T2207/20081G06T2207/20221G06N3/045Y02T10/40
Inventor 侯春萍夏晗杨阳莫晓蕾徐金辰
Owner TIANJIN UNIV
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products