Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Monocular scene depth prediction method based on deep learning

A technology of scene depth and deep learning, which is applied in the field of computer vision and image processing, can solve the problems that the image generation model does not have scaling, the depth model is not completely differentiable, and the gradient computability is lost, so as to alleviate the problem of gradient disappearance and improve The transmission of information and gradients, and the effect of enhancing feature propagation

Active Publication Date: 2020-11-06
SOUTHEAST UNIV
View PDF3 Cites 7 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0005] 3) The method based on the depth information restored by the sparse depth map has the problem of discontinuous edge depth;
[0006] 4) The depth model is not completely differentiable, and the computability of the gradient in optimization is lost, making the training suboptimal;
[0007] 5) The image generation model does not have the ability to scale images to large output resolutions;
[0008] 6) The generalization ability of the model is generally restricted by the training data

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Monocular scene depth prediction method based on deep learning
  • Monocular scene depth prediction method based on deep learning
  • Monocular scene depth prediction method based on deep learning

Examples

Experimental program
Comparison scheme
Effect test

Embodiment 1

[0040] Embodiment 1: as figure 1 As shown, a monocular scene depth prediction method based on deep learning, using calibrated color image pairs as training input, using DenseNet convolution module to improve the network architecture of the encoder part, in the multi-dimensional matching of binocular stereo and image smoothing The loss constraint is strengthened at each level, the occlusion problem is improved by post-processing, and the depth prediction effect of the monocular image is generally improved. The method includes the following steps:

[0041] Step 1: Preprocessing operation, resize the high-resolution binocular color image pair to 256x512, and perform random flip and contrast transformation on the unified image pair to increase the amount of input data. Input into the encoder of the convolutional network.

[0042] Step 2: In the network encoder part, use the DenseNet convolution module to extract visual features, improve the transmission of information and gradien...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses a monocular scene depth prediction method based on deep learning. The method is suitable for monocular pictures or videos, a calibrated binocular color image pair is used for training a depth prediction model, a DenseNet convolution module is used for extracting a feature space in a network architecture, and dense blocks and transition layers in the feature space are used,so that each layer in a network is directly connected with a front layer of the network, and repeated utilization of features is realized; binocular matching loss improvement, a depth prediction problem is regarded as an image reconstruction problem; the method comprises the steps of sampling an input left viewpoint color image and a disparity image to generate a virtual color image and a disparity image, and restraining the consistency of the generated virtual view and a correspondingly input right viewpoint image on an RGB level and a disparity level by utilizing a stereo matching algorithmof a binocular image pair, thereby obtaining a better depth; the depth smooth loss is improved, the high-quality dense depth map can be generated, the artifact problem caused by shielding in monocularscene depth prediction is effectively solved, and the 2D-to-3D conversion requirements of multiple indoor and outdoor real scenes can be met.

Description

technical field [0001] The invention relates to a monocular scene depth prediction method based on deep learning, belonging to the fields of computer vision and image processing. Background technique [0002] Monocular depth prediction is a research topic that has attracted much attention in computer vision, and has a wide range of application values ​​in areas such as autonomous driving, VR game production, and film and television production. However, there are still many unsolved problems in this field, such as: [0003] 1) The process of using radar laser to collect depth data is costly and greatly affected by weather; [0004] 2) The generated depth image produces artifacts and smears due to light shadows or occlusions in the original image; [0005] 3) The method based on the depth information restored by the sparse depth map has the problem of discontinuous edge depth; [0006] 4) The depth model is not completely differentiable, and the computability of the gradien...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(China)
IPC IPC(8): G06T7/55G06T7/13G06N3/04
CPCG06T7/55G06T7/13G06N3/045
Inventor 姚莉缪静
Owner SOUTHEAST UNIV
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products