Unlock instant, AI-driven research and patent intelligence for your innovation.

Robot operation method based on visual and auditory fusion

An operation method and robot technology, applied in the fields of instruments, manipulators, computer parts, etc., can solve problems such as inability to verify system functions without limit, inability to accurately locate targets, complex and changeable task requirements, etc., to achieve processing capabilities and improve interaction. Sex and intelligence effects

Active Publication Date: 2022-02-11
北京具身智能科技有限公司
View PDF6 Cites 3 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

The current embodied intelligent robot system is mainly limited to the simulation environment. For the robot to carry out various simulation experiments in the simulation environment, the ideality of each equipment greatly avoids the cost of data collection and ensures the stability of the entire robot experiment. , so it helps the whole experimental process to go smoothly, but for the real-world robot learning exploration, there are huge physical limitations, and there are few applications of multi-modal information on robots.
When faced with multiple scene targets, human natural language often contains high-level semantic information. For this, robots can only realize simple language information processing and feedback, and cannot accurately locate the target
Moreover, in the real physical environment, there are complex and changeable task requirements, and relying on single-modal information is not enough to complete all tasks. For multiple modal information, the existing technology lacks the ability to realize the same robot to collect different modal processing system
For real physical systems, existing methods cannot verify their system functions without limit

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Robot operation method based on visual and auditory fusion
  • Robot operation method based on visual and auditory fusion
  • Robot operation method based on visual and auditory fusion

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0059] The technical solutions in the embodiments of the present invention will be apparent from the embodiments of the present invention, as will be described, as will be described herein, not all of the embodiments of the invention, not all of the embodiments of the invention. Based on the embodiments of the invention, all other embodiments obtained by those of ordinary skill in the art are in the range of the protection of the present invention.

[0060] like Figure 1-3 As shown, a robot operation method based on visual audio-integrated fusion according to an embodiment of the present invention, including the following steps:

[0061] S1 Generate Visual Scene Images: Scene map of the physical operation platform, the scene diagram is divided into three types of scene maps from the location of the object, the properties of the object, and the relationship between other specific objects into three types of scene maps, using data enhancement methods, three types of scene maps Cut, ...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses a robot operation method based on visual and auditory fusion. The robot operation method comprises the following steps of S1, generating a visual scene image; S2, generating a natural language operation instruction; S3, constructing an auditory data set; S4, constructing an indication expression model; S5, constructing an auditory sense prediction model; S6, constructing an auditory recognizer; S7, constructing an operation model; and S8, repeating the steps S5-S7, and generating the action sequences of the multiple mechanical arms circularly. According to the robot operation method, the visual positioning model and the auditory detection model are constructed and integrated on the mechanical arm operation model, so that a robot system can obtain the visual information, understand that the advanced semantic information in a natural language is positioned to a target with an indication expression relation, and distinguish an object which is difficult to distinguish visually by means of the auditory information; the corresponding operations are executed, so that the multi-modal data processing capacity of the robot and the autonomous operation and environment interaction capacity of the robot are achieved, and the interactivity and intelligence of the multi-modal robot system are improved.

Description

Technical field [0001] The present invention relates to the field of robotic control, and in particular, to a robot operating method based on visual hearing fusion. Background technique [0002] In recent years, with the continuous development of artificial intelligence technology, various advanced algorithms are integrated into the robot, making robots more intelligent. More and more industries can see the service robot to produce life, such as medical robots, hotel rooms, logistics transport robots, etc., which also led to a higher demand for robots, and began to gradually asking robots. Understand the scene of the real world, even able to complete the task according to human natural language instructions. However, human natural language contains uncertainty. The real scene is also complex and variable. Only single modal data can only make the robot to complete orientation, for different target operation tasks, require specific sensors to collect More information Copies differe...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(China)
IPC IPC(8): B25J9/16G06V20/10G06V10/80G06V10/764G06V10/82G06K9/62
CPCB25J9/1694G06N3/045G06F18/241G06F18/253
Inventor 袁小虎王业飞
Owner 北京具身智能科技有限公司