Vehicle detection method based on monocular vision and laser radar fusion

A laser radar and vehicle detection technology, applied in character and pattern recognition, instruments, computer components, etc., can solve the problems of vehicle position and laser radar missed detection, eliminate overlapping detection results, suppress sample imbalance, and reduce the effect of time

Pending Publication Date: 2020-06-16
TONGJI UNIV
View PDF3 Cites 21 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0005] The purpose of the present invention is to provide a vehicle detection method based on the fusion of monocular vision and laser radar in order to overcome the defects of the above-mentioned prior art, which helps to solve the problem that monocular vision is difficult to effectively estimate the vehicle position and laser radar may be due to The problem of missed detection caused by sparse long-distance point clouds further improves the effect of 3D target detection on vehicles

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Vehicle detection method based on monocular vision and laser radar fusion
  • Vehicle detection method based on monocular vision and laser radar fusion
  • Vehicle detection method based on monocular vision and laser radar fusion

Examples

Experimental program
Comparison scheme
Effect test

Embodiment

[0049] A vehicle detection method based on monocular vision and lidar fusion, including: using a feature pyramid network to extract image feature maps; using a VoxelNet network to obtain point cloud feature maps and extract 3D candidate regions; extract point clouds of regions of interest based on candidate regions Features and image features; use the pre-fusion strategy to fuse image features and point cloud features to obtain fusion features; use fusion features to estimate target categories and 3D bounding boxes; use non-maximum value suppression method to remove overlapping redundant bounding boxes and other steps, the overall process Such as figure 2 shown, including:

[0050] Step 1: Normalize the input image. First, calculate the mean value of the training set image on the R / G / B three channels. During training and detection, it is necessary to put the image on the R / G / B three channels The pixel values ​​of the mean are subtracted respectively;

[0051] Step 2: For th...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention relates to a vehicle detection method based on monocular vision and laser radar fusion. The method comprises the following steps of S1, obtaining an image feature map; S2, obtaining a point cloud feature map; S3, respectively extracting a point cloud feature vector flidar and an image feature vector fRGB from the point cloud feature map and the image feature map; S4, performing feature fusion on the point cloud feature vector flidar and the image feature vector fRGB to obtain a fusion feature fL; S5, obtaining a 3D bounding box of the vehicle according to the fusion feature fL, and obtaining corresponding category parameters; S6, removing the overlapped 3D bounding boxes to obtain the final 3D bounding boxes to complete vehicle detection. Compared with the prior art, the method has the advantages that the problems that the position of the vehicle is difficult to effectively estimate through monocular vision and missing detection is possibly caused by laser radar due to long-distance point cloud sparseness are solved, and the three-dimensional target detection effect on the vehicle is further improved.

Description

technical field [0001] The invention relates to the field of autonomous driving environment perception, in particular to a vehicle detection method based on fusion of monocular vision and laser radar. Background technique [0002] Vehicle detection is an essential part of the autonomous driving environment perception system, and target detection is also a basic problem in computer vision. Although researchers have made tremendous progress in this field in recent years, it is still a major challenge to develop a high-accuracy, high-efficiency, high-robust object detection system that can be used for autonomous driving. The detection of the vehicle is realized through 3D target detection. The output of the 3D target detection is a 3D bounding box, which contains the position and attitude information of the target vehicle in the 3D environment. With this information, the automatic driving decision-making system can make further driving decisions. , which is more important for ...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
Patent Type & Authority Applications(China)
IPC IPC(8): G06K9/00G06K9/62G06K9/32
CPCG06V20/584G06V10/25G06V2201/08G06F18/213G06F18/214G06F18/24G06F18/253
Inventor 张立军孟德建黄露莹张状
Owner TONGJI UNIV
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products