Unlock instant, AI-driven research and patent intelligence for your innovation.

Three-dimensional model classification and retrieval method based on visual saliency information sharing

A 3D model and information sharing technology, which is applied in neural learning methods, biological neural network models, character and pattern recognition, etc., can solve the problems of large amount of information in 3D models, high time and space complexity of 3D model classification and retrieval tasks, etc. , to achieve the effect of comprehensive description, increasing scientificity and accuracy, increasing flexibility and stability

Inactive Publication Date: 2020-06-05
TIANJIN UNIV
View PDF3 Cites 9 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0006] 1) Due to the large amount of 3D model information, the 3D model classification and retrieval tasks have high time and space complexity;

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Three-dimensional model classification and retrieval method based on visual saliency information sharing
  • Three-dimensional model classification and retrieval method based on visual saliency information sharing
  • Three-dimensional model classification and retrieval method based on visual saliency information sharing

Examples

Experimental program
Comparison scheme
Effect test

Embodiment 1

[0031] A 3D model classification and retrieval method based on visual saliency information sharing, see figure 1 , mainly includes three parts: one is attention-based view weight calculation; the other is view attention pooling; the final shape descriptor generation, the specific implementation steps are as follows:

[0032] 101: Given a 3D model, extract a view every 30 degrees around the Z-axis of the 3D model, and extract the feature descriptor of each virtual image through a deep convolutional neural network;

[0033] 102: The feature descriptor is used as the input of the visual saliency branch, the weight of the view is generated through the first LSTM module and the soft attention mechanism, and the feature descriptor of the visual saliency branch is generated through the second LSTM module;

[0034] 103: Also use the feature descriptor as the input of the MVCNN branch, use the view weight to guide the fusion of visual information in the MVCNN model, and then obtain the...

Embodiment 2

[0055] Combined with the following network structure, figure 1 , figure 2 The scheme in Example 1 is further introduced, see the following description for details:

[0056] For extracting the first feature descriptor, the present invention takes the Z axis as the center of rotation, samples the perspective of the 3D model at an interval of 30°, and extracts the feature descriptor of the view through a mature deep convolutional neural network, as follows:

[0057] 1. Normalize each 3D model with NPCA (3D Principal Component Analysis) method. Then, a visualization tool developed by OpenGL acts as a human observer to extract a view every 30 degrees around the Z-axis direction of each 3D model. 12 views are extracted to represent the visual and structural information of the 3D model. Thus, these views can be seen as a sequence of images v 1 ,v 2 ,...v 12 , which is very important for the network structure of the present invention.

[0058] 2. Use the network structure of C...

Embodiment 3

[0071] Below in conjunction with specific example, the scheme in embodiment 1 and 2 is carried out feasibility verification, see the following description for details:

[0072] The database in the embodiment of the present invention is based on ModelNet40 and ShapeNetCore55. ModelNet40 is a subset of ModelNet that contains 12,311 CAD models grouped into 40 categories. The model has been manually cleaned, but the pose has not been normalized. The ModelNet40 models used in the embodiment of the present invention are all in *.off format. ShapeNetCore55 is a subset of ShapeNet, which contains 55 categories and about 51,300 3D models. Each category is subdivided into several subcategories, including 70% training set, 10% validation set, and 20% test set. The ShapeNetCore55 model used in the embodiment of the present invention is in *.obj format.

[0073] The following table shows the accuracy of different parts of the network in the ModelNet40 data set for classification experime...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses a three-dimensional model classification and retrieval method based on visual saliency information sharing, and the method comprises the steps: extracting a view every 30 degrees in the Z-axis direction of a three-dimensional model, and extracting a feature descriptor of each virtual image through a deep convolutional neural network; taking the feature descriptor as the input of a visual saliency branch, generating the weight of a view through a first LSTM module and a soft attention mechanism, and generating the feature descriptor of the visual saliency branch througha second LSTM module; taking the feature descriptor as an input of an MVCNN branch, guiding visual information fusion in an MVCNN module by applying a view weight, and obtaining the feature descriptorof the MVCNN branch through a CNN; connecting the descriptors of the two branches in series, carrying out decision making judgment through a full connection layer and a softmax layer, carrying out classification, executing similarity measurement, and carrying out retrieval. Based on a view convolutional neural network and two branches of visual saliency, the feature descriptors are fused to generate the feature descriptors for 3D shape classification and retrieval.

Description

technical field [0001] The invention relates to the fields of three-dimensional model feature extraction, three-dimensional model classification and retrieval, and in particular to a three-dimensional model classification and retrieval method based on visual saliency information sharing. Background technique [0002] In recent years, with the application of 3D technology in the film and television industry, people can see 3D models almost anywhere, so it is natural and reasonable to explore more efficient methods to learn the representation of 3D models. In addition, with the development of computer vision and 3D reconstruction technology, 3D shape recognition has become a basic task of shape analysis, which is the most critical technology for processing and analyzing 3D data. Benefiting from the use of powerful deep learning neural networks and large-scale labeled 3D shape collections, various deep networks for 3D shape recognition have been investigated. Generally, 3D sha...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
IPC IPC(8): G06K9/62G06N3/04G06N3/08
CPCG06N3/08G06N3/044G06N3/045G06F18/24317G06F18/253G06F18/29
Inventor 聂为之王亚屈露
Owner TIANJIN UNIV