Check patentability & draft patents in minutes with Patsnap Eureka AI!

Natural language navigation method based on radar and vision multi-mode fusion

A multi-modal, radar technology, applied in the fields of natural language processing, computer vision, and robot navigation, it can solve problems such as unobservable obstacles, and achieve the effect of improving obstacle avoidance ability and increasing success rate.

Active Publication Date: 2021-07-23
TSINGHUA UNIV
View PDF7 Cites 3 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

Most of the current researchers are improving the accuracy of robot visual language navigation in the simulation environment, but in the real scene, due to the complexity of the environment, the robot often encounters obstacles during the navigation process, relying only on vision and depth information , the robot is often unable to avoid obstacles
Moreover, since the binocular camera is usually installed above the robot, the robot cannot observe the obstacles under its feet

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Natural language navigation method based on radar and vision multi-mode fusion
  • Natural language navigation method based on radar and vision multi-mode fusion
  • Natural language navigation method based on radar and vision multi-mode fusion

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0022] The present invention proposes a embodied language navigation method based on multimodal fusion of radar and vision. The present invention will be further described in detail below in conjunction with the accompanying drawings and specific implementation methods.

[0023] The present invention proposes a embodied language navigation method based on multimodal fusion of radar and vision. The overall process is as follows figure 1 shown, including the following steps:

[0024] 1) Install laser radar on the robot with binocular camera (this embodiment adopts UST-10LX 2D laser radar of HOKUYO company), the laser radar is installed in front of the robot, after the installation is completed, the robot is used as the navigation task robot;

[0025] 2) Construct a multimodal fusion neural network model; the neural network model includes five subnetworks, namely: language encoding subnetwork, RGB image encoding subnetwork, depth image encoding subnetwork, radar information enco...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention provides a natural language navigation method based on radar and vision multi-mode fusion and belongs to the fields of robot navigation, natural language processing and computer vision. The method comprises the following steps of: firstly, installing a laser radar on a robot with a binocular camera, and training a multi-modal fusion neural network model by using the robot; selecting any real scene, issuing a natural language navigation instruction to the robot, and converting the natural language navigation instruction into a corresponding semantic vector; respectively converting RGB images, depth images and radar information acquired by the robot at each moment into corresponding features; performing feature fusion on the semantic vector, a RGB image feature and a depth image feature to obtain an action feature at the current moment; and after correcting the action feature by using radar features, enabling the neural network model finally to output the action of the robot at the current moment, and enabling the robot to execute the action until a navigation task is completed. According to the natural language navigation method based on the radar and vision multi-mode fusion of the invention, the robot can have good perception ability to the real environment, and the efficiency of obstacle avoidance navigation is improved.

Description

technical field [0001] The invention relates to the fields of robot navigation, natural language processing and computer vision, and is a embodied language navigation method based on multimodal fusion of radar and vision. Background technique [0002] The use of natural language for human-computer interaction with computers has long been pursued because of its important theoretical implications as well as its obvious practical implications. Human beings can use their natural language to use computers without spending a lot of time learning various complex computer languages. At present, using natural language to control robots for autonomous navigation is gradually becoming a research hotspot. Researchers hope that in the future, robots can be controlled by natural language to complete navigation tasks, and navigation tasks are also the basis for other complex tasks of robots, which is of great significance to the development of artificial intelligence. [0003] Visual lan...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(China)
IPC IPC(8): G01S13/86G01S13/88G01S13/93G01C11/00G01C21/00G06K9/00G06K9/46G06K9/62G06N3/04G06N3/08
CPCG01S13/867G01S13/881G01S13/93G01C21/005G01C11/00G06N3/084G06V20/10G06V10/44G06N3/047G06F18/241G06F18/2415
Inventor 刘华平周方波袁小虎赵怀林
Owner TSINGHUA UNIV
Features
  • R&D
  • Intellectual Property
  • Life Sciences
  • Materials
  • Tech Scout
Why Patsnap Eureka
  • Unparalleled Data Quality
  • Higher Quality Content
  • 60% Fewer Hallucinations
Social media
Patsnap Eureka Blog
Learn More