Identification method of PointNet for complex scene

A recognition method and complex scene technology, applied in the field of optimizing PointNet's recognition of complex scenes, can solve problems such as objects that cannot be parsed out of details

Inactive Publication Date: 2019-12-31
HANGZHOU DIANZI UNIV
View PDF5 Cites 5 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

Unobtrusive class scenes contain objects of arbitrary size. These details are also important for the recognition of complex scenes. Objects with fine details may not be resolved without the aid of prior information.

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Identification method of PointNet for complex scene
  • Identification method of PointNet for complex scene
  • Identification method of PointNet for complex scene

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0017] The present invention will be further described below in conjunction with drawings and embodiments.

[0018] Such as figure 1 As shown, the present invention realizes a kind of processing 3D recognition tasks in three-dimensional space, including tasks such as object classification, partial segmentation and semantic segmentation, and the specific implementation steps are as follows:

[0019] S1. Input data: a three-dimensional point cloud (n*3) of n points of point cloud data is used as input.

[0020] S2. Predict an effective transformation matrix through the mini-network (T-net), and directly apply this transformation to the coordinates of the input points. Change the input, adjust the point cloud (unordered vector) in the space, and rotate it to an angle that is more conducive to segmentation. Input the point cloud data, first perform affine transformation with T-Net, the specific performance is that the original data passes through a 3D space transformation matrix...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention discloses an identification method of PointNet for complex scene. According to the method, a pyramid pooling module is used for optimizing the recognition capability of Point Net in a complex scene. The feature vector obtained after Point Net processing is accessed to a pyramid pooling module. The pyramid pooling module can improve the performance of open vocabulary object and filling identification in complex scene analysis. The capability of context aggregation based on different regions with global text information is leveraged. The global prior representation is effective forgenerating a high-quality result in a scene analysis task; for a complex scene, the pyramid parsing module provides good description for overall scene interpretation, the knowledge graph depends on prior information of scene contexts, and the pyramid pooling module can aggregate context information of different regions, so that the capability of obtaining global information is improved. Therefore, due to the addition of the pyramid module, tasks such as object classification, partial segmentation, semantic segmentation and the like in a three-dimensional space can have higher accuracy.

Description

technical field [0001] The invention belongs to the field of image retrieval and relates to a recognition method for optimizing PointNet for complex scenes. Background technique [0002] With the rapid development of the Internet and deep learning, we have done a lot of research on two-dimensional images, but for 3D point clouds, it is still a challenge for us. 3D point cloud data is an unordered set of points. Typical convolution structures require a highly regular input data format, so we need to convert these data into regular 3D voxel grids or image collections (such as views), and then Feed it to a deep network architecture. However, this data representation transformation makes the resulting data unnecessarily large. It also introduces quantization artifacts that can obscure the natural invariance of the data. The PointNet network directly takes point clouds as input and outputs class labels for the entire input or per point segment / part labels for each point of the...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
IPC IPC(8): G06K9/00G06K9/34G06K9/62
CPCG06V20/00G06V10/267G06F18/24
Inventor 颜成钢郭凡锋孙垚棋张继勇张勇东
Owner HANGZHOU DIANZI UNIV
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products