Multi-source image fusion and feature extraction algorithm based on depth learning

A technology of feature extraction and deep learning, which is applied in computing, computer parts, character and pattern recognition, etc., can solve the problems of ignoring feature learning and optimization, limiting the scope of practical application, and not being able to represent the topology of 3D models, so as to improve information Effects of Missing, Improving Accuracy, and Improving Accuracy and Efficiency

Inactive Publication Date: 2019-02-05
TIANJIN UNIV +1
View PDF2 Cites 12 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0005] 1. The view itself cannot represent the topology of the target 3D model, and the classification of 3D models based on the view has certain limitations
[0006] 2. How to select the optimal view affects the accuracy of the final target classification to a certain extent, and there is no relatively perfect method for the selection of the optimal view.
[0007] The main challenges currently faced in the field of view-based 3D model retrieval are [10] : Most methods focus on similarity calculation and model structure representation, while ignoring the learning and optimization of features, which limits the scope of practical applications

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Multi-source image fusion and feature extraction algorithm based on depth learning
  • Multi-source image fusion and feature extraction algorithm based on depth learning
  • Multi-source image fusion and feature extraction algorithm based on depth learning

Examples

Experimental program
Comparison scheme
Effect test

Embodiment 1

[0031] A multi-source image fusion and feature extraction algorithm based on deep learning, see figure 1 , the method includes the following steps:

[0032] 101: Place each 3D model in the database in a virtual dodecahedron, place the virtual camera on twenty evenly distributed vertices of the dodecahedron, and view the original object from the viewpoint of the 3D space Take a virtual photo to get 20 views of a single target to form a multi-view model database;

[0033] 102: Divide the multi-view model database into training set, test set and verification set according to the ratio of 7:2:1, use the hidden variable of view pose label to redefine the loss function, and minimize the loss function through the backpropagation algorithm;

[0034] 103: After minimizing the loss function, the last layer of the neural network outputs multiple views of a single target through softmax cascade, and the score of the category under the constraint of the candidate view pose label.

[0035...

Embodiment 2

[0039] The scheme in embodiment 1 is further introduced below in conjunction with specific examples and calculation formulas, see the following description for details:

[0040] 201: ModelNet40 [11] Each 3D model in the database is placed in a virtual dodecahedron, the virtual camera is placed on the 20 vertices of the dodecahedron, and the original object is viewed from these 20 viewpoints uniformly distributed in the 3D space By taking virtual photos, you can get twenty views of a single target;

[0041] Wherein, the above step 201 mainly includes:

[0042] A set of viewpoints are predefined, and the viewpoints are the viewpoints of the object to be observed. Let M be the number of predefined viewpoints. In the embodiment of the present invention, M is taken as 20. Place the virtual camera on the 20 vertices of the dodecahedron containing the target. The dodecahedron is the regular polyhedron with the largest number of vertices, and its viewpoints are completely evenly dis...

Embodiment 3

[0076] Below in conjunction with concrete test, the scheme in embodiment 1 and 2 is carried out feasibility verification, see the following description for details:

[0077] image 3 The scheme in the present embodiment has been carried out feasibility verification, adopts recall rate-precision rate to measure the performance of this method, it uses recall rate (Recall) and precision rate (Precision) as x-axis and y-axis respectively, It can be obtained according to the following formula:

[0078]

[0079] Among them, Recall is the recall rate, N z For the number of correctly retrieved targets, C r is the number of all related targets.

[0080]

[0081] Among them, Precision is the precision rate, C all is the number of all retrieved targets.

[0082] Generally speaking, the larger the area enclosed by the recall-precision rate curve and the coordinate axis, the better the performance of the algorithm. Depend on image 3 It can be seen that the area enclosed by th...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention discloses a multi-source image fusion and feature extraction algorithm based on depth learning, comprises placing each three-dimensional model in a virtual dodecahedron, arranging a virtual camera on twenty vertices of the dodecahedron, taking virtual photos of the original object from the viewpoints of three-dimensional space, and obtaining twenty views of a single object to form amulti-view model database; The multi-view model database is divided into training set, test set and verification set according to the ratio of 7: 2: 1. The loss function is redefined by using the hidden variable of view posture label, and the loss function is minimized by back propagation algorithm. After minimizing the loss function, the last layer of the neural network outputs multiple views ofa single object through a softmax cascade, and scores of the categories under the constraint of the candidate view gesture tags. The invention avoids the dependence on the space where the feature is located, and improves the accuracy of the target classification.

Description

technical field [0001] The present invention relates to the field of multi-view target classification, in particular to a multi-source image fusion and feature extraction algorithm based on deep learning. Background technique [0002] With digitization, display technology and 3D modeling [1] With the increasing maturity of technologies such as 3D models, the goal of 3D models has shown explosive growth. How to efficiently analyze and retrieve this goal has become an important research problem. [2] . while the target class [3] As an important step in 3D model retrieval, it can greatly affect the speed and accuracy of classification. Restricting the scope of retrieval by category can not only improve the efficiency of retrieval, but also improve the accuracy of retrieval. The target classification technology refers to the classification of a part of the marked target (such as: image, video, 3D model, etc. [4] ) as a training sample, it is trained by algorithms such as deep...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
Patent Type & Authority Applications(China)
IPC IPC(8): G06K9/62
CPCG06F18/253G06F18/254
Inventor 周河宇韦莎程雨航王伟忠刘安安聂为之苏育挺
Owner TIANJIN UNIV
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products