Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Triad acquisition method and device based on video learning and text learning

An acquisition method and triplet technology, applied in the field of artificial intelligence, can solve problems such as limiting the autonomous learning ability of robots, low accuracy and efficiency of robots, and inability of robots to determine the connection between objects, achieving high accuracy and good autonomous learning. , the effect of improving the accuracy

Active Publication Date: 2020-11-17
GUANGDONG UNIV OF TECH
View PDF3 Cites 0 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0004] The purpose of the present invention is to at least solve one of the above-mentioned technical defects, especially in the prior art, the robot cannot judge the connection between objects and the connection between actions and objects, thus limiting the autonomous learning ability of the robot, making the robot perform The technical defect of lower accuracy and efficiency of action

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Triad acquisition method and device based on video learning and text learning
  • Triad acquisition method and device based on video learning and text learning
  • Triad acquisition method and device based on video learning and text learning

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0041] The following will clearly and completely describe the technical solutions in the embodiments of the present invention with reference to the accompanying drawings in the embodiments of the present invention. Obviously, the described embodiments are only some, not all, embodiments of the present invention. Based on the embodiments of the present invention, all other embodiments obtained by persons of ordinary skill in the art without making creative efforts belong to the protection scope of the present invention.

[0042] Those skilled in the art will understand that unless otherwise stated, the singular forms "a", "an", "said" and "the" used herein may also include plural forms. It should be further understood that the word "comprising" used in the specification of the present application refers to the presence of the features, integers, steps, operations, elements and / or components, but does not exclude the presence or addition of one or more other features, Integers, ...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention provides a triple acquisition method and device based on video learning and text learning, a co-fusion robot visual programming and online cooperation system and a robot, and the methodcomprises the steps: firstly learning a long video to acquire a main body object, an acceptor object and an action type corresponding to each video segment, and forming a corresponding operation instruction triple; training a natural language model by utilizing the text triple positive and negative sample data set, inputting text information formed by the operation instruction triple into the natural language model to obtain a prediction triple, and correcting the operation instruction triple output by the long video through the prediction triple so as to realize multi-modal fusion; moreover,the accuracy of the target triad obtained through video learning and text learning is relatively high, the accuracy of identifying the long video operation instruction by the robot can be effectivelyimproved, better autonomous learning of the robot is facilitated, and the learning efficiency is improved.

Description

technical field [0001] The present invention relates to the technical field of artificial intelligence, and in particular to a triplet acquisition method and device based on video learning and text learning, a visual programming and online collaboration system for a fusion robot, and a robot. Background technique [0002] With the continuous development of robot technology, the autonomous learning ability of robots is also constantly improving; for example, in the prior art, long videos are usually used to learn operation instruction triplets, and the robot divides the input long videos into video segments according to actions, Then identify the position and type of the main object and the recipient object in each video segment, as well as the action between the main object and the recipient object, thus forming a triplet of operation instructions, and then the robot executes the corresponding operation instructions to complete autonomous learning process, and can make diffe...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(China)
IPC IPC(8): G06K9/00G06K9/62G06N3/04G06N3/08
CPCG06N3/08G06V20/10G06V20/40G06N3/048G06N3/045G06F18/241Y02D10/00
Inventor 黄可思刘文印陈俊洪朱展模王思涵
Owner GUANGDONG UNIV OF TECH
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products