Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Industrial man-machine interaction system and method based on vision and inertial navigation positioning

An inertial positioning, human-machine technology, applied in manipulators, manufacturing tools, program-controlled manipulators, etc., can solve problems such as inability to interact with humans and move around with people, complex image processing processes, and space constraints.

Active Publication Date: 2021-04-27
SHENYANG INST OF AUTOMATION - CHINESE ACAD OF SCI
View PDF9 Cites 7 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

The human-computer interaction process of the existing scheme is still limited by space, and it cannot make the human-computer interaction free from space constraints and walk with people, and the recognition image processing process is complicated

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Industrial man-machine interaction system and method based on vision and inertial navigation positioning
  • Industrial man-machine interaction system and method based on vision and inertial navigation positioning
  • Industrial man-machine interaction system and method based on vision and inertial navigation positioning

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0055] The present invention will be described in detail below in conjunction with the accompanying drawings.

[0056] The industrial human-computer interaction system based on vision and inertial navigation positioning includes controlled equipment, scene image acquisition system and auxiliary positioning equipment worn on the operator's arm, palm and fingertips.

[0057] In this system, the scenes of human-computer interaction can be divided into two categories: non-interactive workspace and interactive workspace. During the working process in the interactive workspace, the operator will have direct or indirect contact with the robotic arm. The main problem that threatens people and safety in the process of human-computer interaction is the collision between the operator and the controlled equipment. The main task of the system during operation is to predict the impending collision. For two different scenarios, this system divides the problem of ensuring human-machine safet...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention relates to an industrial man-machine interaction system and method based on vision and inertial navigation positioning. The system comprises at least one manipulator, at least one mechanical arm, at least four industrial cameras, at least four cameras, at least one set of somatosensory wearable equipment and at least one computer, wherein the manipulators are used for grabbing and machining target workpieces; the bottom of each mechanical arm is provided with a base, and each manipulator is installed at the tail end of one mechanical arm; the industrial cameras are used for acquiring color images of arms, finger joints and the manipulators in a working area of an operator; the cameras are used for stereoscopically positioning the mechanical arms and the manipulators in a space; the somatosensory wearable equipment is worn at the arm position of the operator and used for acquiring displacement, speed and acceleration information in the moving direction; and the computers are used for acquiring information and judging whether the operator collides with the mechanical arms and the manipulators or not by acquiring the information of the wearable equipment and image information of the cameras in a small view field of the working area of the operator.

Description

technical field [0001] The invention relates to an industrial human-computer interaction system and method based on vision and inertial navigation positioning, belonging to the fields of human-computer interaction and industrial measurement. Background technique [0002] At present, there are many human-computer interaction methods to ensure human-machine safety in industrial sites, including using hand-held controllers to ensure that people are within the safe range, guardrails to ensure personal safety, and operators to ensure personal safety outside the working range of robots. These human-computer interaction methods are large. Most of them are serial, and there is no real contact between man and machine. In the existing human-computer interaction system, the main work of the operator is loading the robot's working materials, supervising the operation of the robot at a safe distance, and inspecting the finished product. The human-computer interaction process of the exis...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(China)
IPC IPC(8): B25J9/16
CPCB25J9/1697B25J9/1676B25J9/1679
Inventor 丛杨侯冬冬唐旭
Owner SHENYANG INST OF AUTOMATION - CHINESE ACAD OF SCI
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products