Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Feature fusion method of multi-modal deep neural network

A feature fusion and multi-modal technology, applied in the field of medical imaging and deep learning, can solve the problems of feature weight distribution analysis and processing, and achieve the effect of maximizing performance

Active Publication Date: 2021-01-29
ZHEJIANG LAB +1
View PDF3 Cites 3 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

However, these methods did not analyze and process the feature weight distribution of each mode, but simply added, superimposed or fused directly

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Feature fusion method of multi-modal deep neural network
  • Feature fusion method of multi-modal deep neural network

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0016] Taking the PET / CT dual mode as an example (that is, when x=2), the present invention will be described in detail in conjunction with the accompanying drawings.

[0017] Such as figure 1 Shown, the inventive method specifically comprises the following steps:

[0018] Step 1: In the dual-branch dual-modal 3D CNN, the two branches correspond to the convolution branch of the PET modality and the convolution branch of the CT modality respectively. For the 3D feature maps output by the nth level of the two 3D convolution branches, the 3D feature maps of the two modalities are superimposed on the channel dimension to obtain a 3D feature map with twice the number of original channels. Then perform average pooling in the three dimensions of depth, height, and width, and compress the three dimensions of depth, height, and width to obtain a one-dimensional vector of a channel dimension. After downsampling and upsampling with a compression ratio of 16:1:16, and using the activati...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses a feature fusion method for a multi-modal deep neural network, and the method comprises the steps of obtaining a channel attention mask between modals in a multi-modal deep three-dimensional CNN through employing a squeeze and excitation (SE) module on a deep learning feature domain, i.e., in all modals, giving more attention to the channels which are remarkably helpful tothe task target, so that the weight distribution of the multi-modal three-dimensional depth feature map on the channels is established explicitly; and then, calculating by utilizing four-dimensional convolution and a Sigmoid activation function to obtain a spatial attention mask between modals, namely, in the three-dimensional feature map of each modal, which positions in the space need to be moreconcerned, so that the spatial correlation of the multi-modal three-dimensional depth feature map is established explicitly, and more attention is given to the positions with important information inmodes, channels and spaces, so that the diagnosis efficiency of the multi-mode intelligent diagnosis system is improved.

Description

technical field [0001] The invention relates to the fields of medical imaging and deep learning, in particular to a feature fusion method of a multimodal deep neural network. Background technique [0002] Existing tumor detection and diagnosis methods are usually realized through medical imaging technology, including planar X-ray imaging, CT, MRI, PET / CT, ultrasound and other modalities, and tissue biopsy is performed on suspicious lesions found in the images. However, due to the heterogeneity of tumors, their properties cannot be fully characterized by a single modality image. For example, on plane X-ray and CT images, the X-ray absorption degree of tumor tissue is characterized; on MRI images, the hydrogen proton density of tumor tissue is characterized; on FDG PET / CT, the It is the activity of tumor tissue to metabolize glucose; on ultrasound images, it is characterized by the reflection degree of tumor tissue to sound waves. Therefore, more and more clinical research b...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
IPC IPC(8): G06K9/62G06N3/04
CPCG06N3/045G06F18/253
Inventor 陈凌朱闻韬张铎申慧李辉叶宏伟王瑶法
Owner ZHEJIANG LAB
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products