Pyramid Transform-based point cloud reconstruction method, device and equipment, and medium

A pyramid and point cloud technology, applied in the field of computer vision, can solve the problem of low reconstruction accuracy of object surface details, and achieve the effect of high sampling efficiency, accurate sampling and high precision

Active Publication Date: 2021-12-31
HUAZHONG UNIV OF SCI & TECH
View PDF15 Cites 5 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0005] Aiming at the defects and improvement needs of the prior art, the present invention provides a point cloud reconstruction method, device, equipment and medium based on a pyramid Transformer, aiming to solve the technical problem of low reconstruction accuracy of object surface details in 3D reconstruction

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Pyramid Transform-based point cloud reconstruction method, device and equipment, and medium
  • Pyramid Transform-based point cloud reconstruction method, device and equipment, and medium
  • Pyramid Transform-based point cloud reconstruction method, device and equipment, and medium

Examples

Experimental program
Comparison scheme
Effect test

Embodiment 1

[0031] refer to figure 1 , combined with figure 2 , the embodiment of the present invention provides a kind of point cloud reconstruction method based on Pyramid Transformer, comprising:

[0032] S1. Obtain multiple views of the target object under different viewing angles, and process each view to obtain corresponding point feature data.

[0033] In this embodiment, in order to obtain more accurate point features, multi-scale global and local information is extracted from the input image. For the local features of points, this program uses the VGG16 network to encode the input image, and outputs the results of different depth convolutional neural networks to obtain multi-scale feature maps. After obtaining the feature maps of these two-dimensional images, use the intrinsic matrix of the camera to project the initial random point cloud onto the feature map, and obtain the coordinates of each point on the feature map. The feature corresponding to the coordinates in the feature...

Embodiment 2

[0060] refer to image 3, the present invention provides a pyramid Transformer-based point cloud reconstruction device 300 provided by an embodiment of the present invention, and the device 300 includes:

[0061] A point feature acquisition module 310, configured to acquire multiple views of the target object under different viewing angles, and process each view to obtain corresponding point feature data;

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention discloses a pyramid Transform-based point cloud reconstruction method, device and equipment, and a medium, and belongs to the field of computer vision. The method comprises the steps that: after point feature data corresponding to each view are obtained, the point feature data are input into a two-channel pyramid network, and point features input in each layer pass through three modules of sampling, a graph neural network and Transformer; when the point feature data passes through the first channel, the updated output of each layer of attention mechanism is input into the next layer, and when the point feature data passes through the second channel, the updated output of each layer of attention mechanism and the updated output of the attention mechanism of the corresponding layer of the first channel located in the same layer as the next layer are spliced and then input into the next layer; and the updated output of the attention mechanism of the bottommost layer of the second channel is converted into a point cloud coordinate to obtain a point cloud result of each view, and an intersection of the point cloud results of each view is taken to represent a target object after point cloud reconstruction. Therefore, the three-dimensional reconstruction precision of the object is higher.

Description

technical field [0001] The invention belongs to the field of computer vision, and more specifically relates to a method, device, equipment and medium for point cloud reconstruction based on a pyramid Transformer. Background technique [0002] The multi-view stereoscopic 3D reconstruction technology reconstructs a 3D object by using multiple input pictures of different perspectives. This technology studies how to obtain the 3D information of the object in space through the 2D information of the picture, and finally obtain the 3D model of the object in space. The traditional 3D reconstruction technology inputs multiple pictures with different perspectives, estimates the camera parameters corresponding to each picture, and then reprojects the objects on different pixels in the pictures to the 3D space, thereby reconstructing the 3D structure of the object. However, the traditional method requires more input images, and the reconstruction effect on reflective objects and objects...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
Patent Type & Authority Applications(China)
IPC IPC(8): G06T17/00G06T7/55G06N3/04G06N3/08
CPCG06T17/00G06T7/55G06N3/04G06N3/08G06T2207/10028G06T2207/20016G06T2207/20081G06T2207/10016G06T2207/20084G06N3/045G06T9/001G06T3/4046G06T9/002
Inventor 刘琼张军杨铀
Owner HUAZHONG UNIV OF SCI & TECH
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products