Embedded attitude learning method capable of carrying out self supervision on the basis of video time-space relationship

A technology of time-space relationship and learning method, applied in the fields of instruments, character and pattern recognition, computer parts, etc., can solve the problems of high cost and time-consuming, and achieve the effect of reducing cost and improving retrieval efficiency.

Inactive Publication Date: 2018-03-16
SHENZHEN WEITESHI TECH
View PDF2 Cites 18 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0004] Aiming at the problem of high cost and time-consuming, the present invention uses spatio-temporal relationship training video to conduct self-supervised learning of pose em

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Embedded attitude learning method capable of carrying out self supervision on the basis of video time-space relationship
  • Embedded attitude learning method capable of carrying out self supervision on the basis of video time-space relationship
  • Embedded attitude learning method capable of carrying out self supervision on the basis of video time-space relationship

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0025] It should be noted that, in the case of no conflict, the embodiments in the present application and the features in the embodiments can be combined with each other. The present invention will be further described in detail below in conjunction with the drawings and specific embodiments.

[0026] figure 1It is a system flowchart of a self-supervised embedding attitude learning method based on the temporal-spatial relationship of videos in the present invention. It mainly includes self-supervised pose embedding: temporal order and spatial layout (1); creating training courses (2); mining repetitive poses (3); network structure (4).

[0027] In supervised training using human annotations, avoid hard examples with ambiguous or even incorrect labels. This kind of data can inhibit convergence and lead to poor results. On the other hand, skipping too many difficult training examples may lead to poor results. Overfitting to a small fraction of easy samples, leading to generali...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention puts forward an embedded attitude learning method capable of carrying out self supervision on the basis of a video time-space relationship. The average value of two auxiliary tasks is used for learning the time-space relationship in a video: one time sequence task learns whether two specific figure images are near or not on an aspect of time, and one spatial arrangement task learns ahuman body appearance model from space to enhance a capability of separating a gesture from a background. On the basis of the learning and mining repeated gesture of a course, training is carried outfrom a simple sample, then, the simple sample is iteratively expanded to a harder sample, and meanwhile, inactive video parts are eliminated. Time-space embedding successfully learns the representative characteristics of human gestures in a self-supervision way. By use of the method, a time-space relationship training video is used for carrying out gesture embedding self-supervision learning, manual annotation is not required, cost is lowered, gesture embedding can capture the visual characteristics of the human gestures, and human gesture estimation and retrieval efficiency can be improved.

Description

technical field [0001] The invention relates to the field of video gesture analysis, in particular to a self-supervised embedding gesture learning method based on the temporal-spatial relationship of the video. Background technique [0002] The ability to recognize human poses is essential for describing actions, and different poses in videos form a visual vocabulary similar to text. In computer vision processing systems, finding similar poses in different videos automatically enables many different applications, such as action recognition or video content retrieval. As an emerging topic, posture analysis has practical development in many fields, such as image search, behavior classification, security monitoring, especially in driverless driving in the transportation field, action recognition in smart homes, and human detection in medical diagnosis. Computer interaction and so on have broad application prospects. According to the needs proposed by video embedding poses, ca...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
IPC IPC(8): G06K9/00G06K9/62
CPCG06V40/23G06F18/214
Inventor 夏春秋
Owner SHENZHEN WEITESHI TECH
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products