Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Monocular vision scene depth estimation method based on deep learning

A monocular vision and scene depth technology, applied in computing, image data processing, instruments, etc., can solve the problems of slow reasoning, expensive acquisition methods, and low accuracy of results, achieving strong constraints, improving computer vision technology, and improving The effect of inference speed

Inactive Publication Date: 2019-10-08
NORTHWESTERN POLYTECHNICAL UNIV
View PDF0 Cites 20 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0004] However, the method of scene depth estimation based on deep learning for monocular images has the following disadvantages: First, the general method of depth estimation for monocular images based on deep learning regards depth restoration as a classification task of image pixels, which requires real Depth data is used to train the network, and this kind of data usually needs to be obtained by lidar, which is not only expensive to obtain but also sparse in data, and poor in practicability; second, the previous method has a single network structure and a relatively simple model, which ultimately leads to low accuracy of the results. Low, and the transferability of the model is poor; third, in order to improve the accuracy of the model, the previous method usually uses a deeper network structure to extract the feature information of the image, resulting in a huge amount of parameters and slow inference speed in practical applications. Poor real-time performance

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Monocular vision scene depth estimation method based on deep learning
  • Monocular vision scene depth estimation method based on deep learning
  • Monocular vision scene depth estimation method based on deep learning

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0035] Now in conjunction with embodiment, accompanying drawing, the present invention will be further described:

[0036] The hardware environment of this experiment is: GPU: Intel Xeon series, memory: 8G, hard disk: 500G mechanical hard disk, independent graphics card: NVIDIA GeForce GTX 1080Ti, 11G; the system environment is Ubuntu 16.0.4; the software environment is python3.6, opencv4.0, Tensorflow.

[0037] In this paper, two sets of experiments have been done on the depth estimation of monocular images. One set is based on the KITTI public data set to verify the accuracy and effectiveness of the invented method; the other set is based on the actual collected monocular image data to verify the The practicality of the method.

[0038] The present invention is specifically implemented as follows:

[0039] Step 1 Construction and training of the network model: using the standard VGG-13 network model, using depth separable convolution to replace the standard convolution in ea...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention relates to a monocular vision scene depth estimation method based on deep learning, and the method comprises the steps of employing a VGG-13 network model, employing a depth separable convolution layer to replace a standard convolution layer so as to reduce the model parameter quantity, and obtaining a network model which can be used for obtaining a parallax image; inputting the monocular image into the trained network model to generate the disparity maps of multiple scales, and generating a single disparity map consistent with the input image in scale by combining multi-scale fusion and disparity map smoothing; and generating a corresponding depth image according to the geometric transformation relationship between the disparity map and the depth map in the multi-view geometry. The method has the beneficial effects that the simple and easily available binocular visible light image is used for training the network model without acquiring the high-cost real depth data, andthe standard convolution is replaced with the depth separable convolution, so that the parameter quantity of the network model can be reduced to one seventh of the previous parameter quantity, and the reasoning speed of the model is increased.

Description

technical field [0001] The invention belongs to the field of three-dimensional reconstruction of computer vision, and relates to a method for estimating the depth of a monocular vision scene based on deep learning. Background technique [0002] Three-dimensional structural information is indispensable information for human beings to observe and understand the environment, understand and analyze the scene. Correctly judging and recognizing the three-dimensional structure of the scene allows the computer to perform tasks such as target positioning and path planning more reasonably and accurately. Therefore, how to reconstruct the three-dimensional structure of the scene from two-dimensional images or video sequences is a major research topic in the field of computer vision. Key points and difficult points. 3D scene reconstruction mainly depends on obtaining the depth map corresponding to the image. At present, there are mainly two ways to obtain depth information. One is to...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(China)
IPC IPC(8): G06T7/50
CPCG06T7/50G06T2207/20081G06T2207/20084G06T2207/20228
Inventor 李晖晖刘浪涛袁翔郭雷刘航
Owner NORTHWESTERN POLYTECHNICAL UNIV
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products