A Human Pose Estimation Method Based on Joint Relationship

A technology of human body posture and joints, applied in computing, computer parts, instruments, etc., can solve problems such as inability to train, a large number of calculations, increase the difficulty of prediction, etc., and achieve the effect of good recognition effect, high computing efficiency, and accurate positioning.

Active Publication Date: 2022-04-05
TONGJI UNIV
View PDF4 Cites 0 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

This approach cannot be trained end-to-end, and the graph convolutional network stage is computationally intensive
[0005] The human body is a non-rigid body, and the rotation of each joint has a great degree of freedom. The free rotation of multiple joints can be superimposed on each other. The joints at the far end of the limbs, such as wrist joints and ankle joints, also have multiple degrees of freedom. The location changes, which increases the difficulty of prediction, but the existing methods do not pay attention to the difference in the detection difficulty of different types of joint points brought about by the special structure of the human body; at the same time, in the multi-person pictures that are closer to the real scene, due to Interaction with people, the diversity of positional relationships and the complexity of life scenes, the position of key points of the human body and body information are all facing serious occlusion problems
The occluded human joints are also part of the human pose, and the existing methods do not take targeted optimization measures for the invisible joints that are more difficult to detect correctly

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • A Human Pose Estimation Method Based on Joint Relationship
  • A Human Pose Estimation Method Based on Joint Relationship
  • A Human Pose Estimation Method Based on Joint Relationship

Examples

Experimental program
Comparison scheme
Effect test

Embodiment

[0063] Such as figure 2 As shown, a kind of human body posture estimation method based on joint relationship provided by the present invention mainly includes the following four steps:

[0064] 1) Construct the joint relationship module to generate supplementary features for auxiliary positioning of difficult joint points, including two sub-modules: the channel-based feature relationship module and the adjacent joint space relationship module;

[0065] 2) Construct a human body pose estimation model based on joint relationships based on the general deep convolutional neural network model;

[0066] 3) Use the marked human body posture data to train the constructed human body posture estimation model based on the joint relationship, and obtain a network model that can better locate limb joints and occluded joints;

[0067] 4) For the input image to be processed, use the human body pose estimation network with joint relationship module trained in step 3) to perform the human bo...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The present invention relates to a human body posture estimation method based on joint relationship, comprising the following steps: S1: building a joint relationship module, including two sub-modules, a channel-based feature relationship module and an adjacent joint space relationship module; S2: building a human body based on joint relationship Pose estimation model; S3: use the labeled human pose data to train the human pose estimation model based on the joint relationship; S4: use the trained human pose estimation model with the joint relationship module to perform the human pose estimation task based on a single image, and obtain the prediction human body posture. Compared with the prior art, the present invention effectively overcomes the problem that it is difficult to detect the position of limb joints with high degrees of freedom such as wrist joints, ankle joints and hidden invisible joints in the image, and the accuracy of human body posture estimation is high.

Description

technical field [0001] The invention relates to the field of human body pose estimation, in particular to a method for estimating human body pose based on joint relationships. Background technique [0002] Human pose estimation is a traditional task in the field of computer vision. Human pose estimation includes the detection of human key points and the generation of human poses. The "key points" in human key point detection refer to important joints such as the top of the human body, shoulders, elbow joints, wrist joints, and ankle joints. The generated human body posture is the complete human skeleton information. With the innovation of computer vision technology, human pose estimation has also gone through a process from manually extracting features to using deep convolutional neural networks as a tool. In recent years, the development of basic deep convolutional neural network structure and performance has also greatly improved the level of extracting human joint featu...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
Patent Type & Authority Patents(China)
IPC IPC(8): G06V40/10G06V10/774G06K9/62G06N3/04
CPCG06N3/045G06F18/214
Inventor 梁爽储港谢驰王颉文
Owner TONGJI UNIV
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products