Unlock instant, AI-driven research and patent intelligence for your innovation.

A robot operation method based on visual and auditory fusion

An operation method and robot technology, applied in the fields of instruments, manipulators, computer parts, etc., can solve problems such as the inability to verify system functions without limit, complex and changeable task requirements, insufficient single-modal information to complete all tasks, etc. Interactivity and intelligence, the effect of realizing processing power

Active Publication Date: 2022-03-29
北京具身智能科技有限公司
View PDF0 Cites 0 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

The current embodied intelligent robot system is mainly limited to the simulation environment. For the robot to carry out various simulation experiments in the simulation environment, the ideality of each equipment greatly avoids the cost of data collection and ensures the stability of the entire robot experiment. , so it helps the whole experimental process to go smoothly, but for the real-world robot learning exploration, there are huge physical limitations, and there are few applications of multi-modal information on robots.
When faced with multiple scene targets, human natural language often contains high-level semantic information. For this, robots can only realize simple language information processing and feedback, and cannot accurately locate the target
Moreover, in the real physical environment, there are complex and changeable task requirements, and relying on single-modal information is not enough to complete all tasks. For multiple modal information, the existing technology lacks the ability to realize the same robot to collect different modal processing system
For real physical systems, existing methods cannot verify their system functions without limit

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • A robot operation method based on visual and auditory fusion
  • A robot operation method based on visual and auditory fusion
  • A robot operation method based on visual and auditory fusion

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0059] The technical solutions in the embodiments of the present invention will be clearly and completely described below. Obviously, the described embodiments are only some of the embodiments of the present invention, but not all of them. All other embodiments obtained by persons of ordinary skill in the art based on the embodiments of the present invention belong to the protection scope of the present invention.

[0060] Such as Figure 1-3 As shown, the robot operation method based on visual and auditory fusion according to the embodiment of the present invention includes the following steps:

[0061] S1 Generate visual scene images: Collect the scene graphs of the physical operation platform. The scene graphs are divided into three types of scene graphs by the positional relationship of objects, the attribute relationship of objects, and the relationship of other specific objects. Using data enhancement methods, the three types of scene graphs Perform cropping, rotation, ...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses a robot operating method based on visual and auditory fusion. The robot operating method includes the following steps: S1 generates visual scene images; S2 generates natural language operation instructions; S3 constructs an auditory data set; S4 constructs an instruction expression model; S5 constructs Auditory prediction model; S6 builds the auditory recognizer; S7 builds the operation model; S8 repeats S5-S7 to generate action sequences of multiple robotic arms in a loop. The robot operation method integrates the visual positioning model and the auditory detection model into the robot arm operation model, so that the robot system can not only obtain visual information, but also understand the advanced semantic information in natural language to locate the target with an instructional expression relationship. Relying on auditory information to distinguish objects that are difficult to distinguish visually and perform corresponding operations, the robot's multi-modal data processing capabilities, the robot's ability to operate autonomously and interact with the environment are realized, and the interactivity and intelligence of the multi-modal robot system are improved.

Description

technical field [0001] The invention relates to the technical field of robot control, in particular to a robot operation method based on visual and auditory fusion. Background technique [0002] In recent years, with the continuous development of artificial intelligence technology, various advanced algorithms have been integrated into robots, making robots more intelligent. More and more industries can see that service robots are used in production and life, such as medical robots, hotel food delivery robots, logistics transportation robots, etc. Understand real-world scenarios, and even be able to complete tasks based on human natural language instructions. However, human natural language contains uncertainty, and the real scene is also complex and changeable. Only single-modal data can only enable the robot to complete simple tasks of orientation. For different target operation tasks, specific sensor collection is required. Additional information addresses different miss...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Patents(China)
IPC IPC(8): B25J9/16G06V20/10G06V10/80G06V10/764G06V10/82G06K9/62
CPCB25J9/1694G06N3/045G06F18/241G06F18/253
Inventor 袁小虎王业飞
Owner 北京具身智能科技有限公司