Unlock instant, AI-driven research and patent intelligence for your innovation.

Multi-modal data fusion method based on compound collaborative structure feature recombination network

A technology of structural features and data fusion, applied in biological neural network models, neural learning methods, neural architectures, etc., can solve semantic asynchrony, does not consider the information interaction between modals, and the modal generalization ability is not strong And other issues

Active Publication Date: 2021-09-10
WUHAN UNIV +3
View PDF3 Cites 6 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

The fusion of feature levels means that different multimodal features are associated in the early stage, and the combined features are pulled into the subsequent unified analysis to provide better information supplementation, but there is a problem of semantic asynchrony between different modal features
For multi-modal feature fusion technology, the existing methods generally only perform simple splicing, dot multiplication, and addition operations on the features of each modality, without considering the information interaction between the modalities, and the semantic gap between features is not considered. Difficulties are not solved enough, and the modal generalization ability is not strong, so it is urgent to explore more effective fusion methods

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Multi-modal data fusion method based on compound collaborative structure feature recombination network
  • Multi-modal data fusion method based on compound collaborative structure feature recombination network
  • Multi-modal data fusion method based on compound collaborative structure feature recombination network

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0075] The technical solutions in the embodiments of the present invention will be clearly and completely described below in conjunction with the accompanying drawings in the present invention. Obviously, the described embodiments are part of the embodiments of the present invention, but not all of them. Based on the embodiments of the present invention, all other embodiments obtained by persons of ordinary skill in the art without making creative efforts belong to the protection scope of the present invention.

[0076] Hereinafter, embodiments of the present invention will be described with reference to the drawings. Such as Figure 1-2 as shown, figure 1 It is a flow chart of the multi-modal data fusion method based on the complex collaborative structural feature recombination network in the embodiment of the present invention. It introduces the preprocessing of image and text raw data and the extraction of deep network features, and the semantic analysis of two heterogeneo...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention provides a multi-modal data fusion method based on a compound collaborative structure feature recombination network. The existing multi-modal data fusion technology is mainly characterized by direct feature fusion, neglects bidirectional interaction between modals, and has the problem of semantic gap between features during multi-modal fusion. In order to solve said problems, the method of the invention comprises the following steps: extracting image and text single-mode features by using a deep neural network to establish an image-text bidirectional interaction attention model based on a transform mechanism; mining feature relation between images and texts, performing multi-modal semantic association, and introducing a compound collaborative structure network to deepen communication of interaction information between modals; performing feature two-way recombination under multi-modal deep fusion, so that image and text semantic space alignment is realized. The invention better adapts to search of complementary information among different modals by a neural network, the understanding and generalization ability of a model to multi-modal semantics is enhanced, and the classification accuracy of a multi-modal feature network is further improved.

Description

technical field [0001] The invention relates to the field of feature extraction in deep learning, in particular to a multimodal data fusion method based on a complex collaborative structure feature recombination network. Background technique [0002] In the field of deep learning, due to the single training and prediction of single-modal data, there will be cases where the data utilization rate is low and the correlation between the attributes of the object of interest cannot be reflected. Therefore, it is necessary to establish the correlation and comprehensive analysis of these data from multiple dimensions, and achieve better results through multi-modal feature fusion. The fusion of feature levels means that different multimodal features are associated in the early stage, and the combined features are pulled into subsequent unified analysis to provide better information supplementation, but there is a problem of semantic asynchrony between different modal features. For m...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(China)
IPC IPC(8): G06K9/62G06N3/04G06N3/08
CPCG06N3/08G06N3/045G06F18/241G06F18/253Y02D10/00
Inventor 秦亮余金沄张敏韩谷静吴文炤赵峰许中平秦旭弘刘开培
Owner WUHAN UNIV
Features
  • R&D
  • Intellectual Property
  • Life Sciences
  • Materials
  • Tech Scout
Why Patsnap Eureka
  • Unparalleled Data Quality
  • Higher Quality Content
  • 60% Fewer Hallucinations
Social media
Patsnap Eureka Blog
Learn More