Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

A pixel-level target location method based on fusion of laser and monocular vision

A monocular vision, target positioning technology, applied in the direction of using optical devices, instruments, measuring devices, etc., can solve the problem of computing resource consumption, object positioning accuracy, effective positioning distance and cost, limited computing power of terminal computing equipment, and difficult objects. positioning and other problems, to achieve the effect of improving autonomous operation ability, ensuring real-time positioning, and high positioning accuracy

Active Publication Date: 2021-11-16
ZHEJIANG UNIV
View PDF9 Cites 0 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

At present, depth sensors such as binocular vision, laser radar, and millimeter-wave radar are usually used in the perception system for object positioning. There are still obvious deficiencies in positioning accuracy, effective positioning distance and cost
Since intelligent autonomous operation equipment is constrained by issues such as cost, volume, and power consumption, the computing power of the terminal computing equipment it carries is relatively limited. Therefore, the object positioning method using the above scheme is expensive, has poor real-time performance, and has limited positioning range. It is difficult to effectively complete the precise positioning of objects under the actual working conditions of the equipment

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • A pixel-level target location method based on fusion of laser and monocular vision
  • A pixel-level target location method based on fusion of laser and monocular vision
  • A pixel-level target location method based on fusion of laser and monocular vision

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0030] In order to make the purpose, technical solution and advantages of the present invention clearer, the present invention will be further described in detail below in conjunction with the accompanying drawings.

[0031] like Figure 1-2 As shown, a pixel-level target positioning method based on laser and monocular vision fusion of the present invention includes the following steps:

[0032] S1. The installation position of the camera ensures that the optical axis of the camera is parallel to the ground. The installation position of the laser ranging module ensures that the line connecting the optical center of the laser ranging module and the optical center of the camera is perpendicular to the ground and makes the vertical distance between the laser ranging module and the monocular camera spacing as small as possible, where O W x W Y W Z W The coordinate system of the laser ranging module is the world coordinate system of this system, O C x C Y C Z C is the camer...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The present invention relates to the field of object three-dimensional positioning methods, in particular to a pixel-level target positioning method based on the fusion of laser and monocular vision. The relative position is calibrated to obtain the calibration parameters. After the monocular camera is used to obtain the environmental image, preprocessing is performed and the pixel coordinates of the target to be located are determined. Then, the actuator is driven by the PID feedback control to drive the laser ranging module to rotate to collect data and perform calculations. Finally, the laser ranging module is driven to accurately hit the ranging light spot to the target position, and then the data is collected to achieve target positioning through calculation. The invention consumes low computing resources, can ensure real-time performance under the condition of limited mobile terminal computing power, has high-performance object positioning capabilities, high precision, and wide range, can improve the operating range of intelligent autonomous operation equipment, and has a certain cost Advantages, lower the threshold of use, and facilitate technology promotion and application popularization.

Description

technical field [0001] The invention relates to the field of three-dimensional object positioning methods, in particular to a pixel-level target positioning method based on laser and monocular vision fusion. Background technique [0002] In recent years, with the development of robot technology and artificial intelligence technology, more and more intelligent autonomous operation equipment has been applied in the fields of social production and life, military and national defense. In social production and life, it is a trend of social development to use intelligent robots to assist or replace humans in some repetitive or service work, which can effectively improve production efficiency and improve the convenience of daily life; in the field of military defense, the use of intelligent robots To assist or replace people to perform specific tasks can effectively improve the efficiency of task execution and ensure the safety of task performers. [0003] Intelligent autonomous o...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Patents(China)
IPC IPC(8): G01B11/00G01B11/02
CPCG01B11/002G01B11/02
Inventor 王滔张雲策葛鸿昌朱世强
Owner ZHEJIANG UNIV
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products