A method for estimating that posture of a three-dimensional human body based on a video stream

A technology of human posture and video streaming, applied in the field of virtual reality, can solve problems such as poor inference effect, difficult learning, and poor practical application effect, and achieve the effects of improving accuracy and real-time performance, reducing nonlinearity, and avoiding cumulative errors

Active Publication Date: 2019-01-25
QINGDAO RES INST OF BEIHANG UNIV
View PDF3 Cites 25 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0006] As mentioned above, the existing 3D human pose estimation has the following technical defects: A. Generally, the 3D coordinates of human joint points are directly output, which is very difficult for the network to learn, because the learning task from the feature space to the 3D pose space is a highly Non-linear learning tasks have high nonlinear shortcomings; B. When performing 3D inference of joint points, the intermediate features of the neural network have not been fully utilized, and it is difficult to combine fea

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • A method for estimating that posture of a three-dimensional human body based on a video stream
  • A method for estimating that posture of a three-dimensional human body based on a video stream
  • A method for estimating that posture of a three-dimensional human body based on a video stream

Examples

Experimental program
Comparison scheme
Effect test

Embodiment 1

[0043] Example 1, such as image 3 As shown, the method of 3D human pose estimation based on video stream is as follows:

[0044] For the first frame of the video, 1) input the two-dimensional image of the current frame, use the hourglass network module to extract the two-dimensional posture of the human body, and generate a heat map of the two-dimensional joint points of the human body in the first frame; The heat map is output to the 3D joint point inference module, and the 2D to 3D spatial mapping is performed to generate the 3D joint point heat map of the human body;

[0045] In the second frame of the video, 1) input the two-dimensional image of the current frame, and use the shallow neural network module to generate a shallow image; Input it to the LSTM module to generate a deep feature map; 3) The deep image feature map generated by the current frame is output to the residual module to generate the heat map of the two-dimensional joint points of the human body in the c...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

A method for estimating that posture of a three-dimensional human body based on a video stream. The method based on depth learning is used to estimate 3D pose of human body in video stream, which avoids many defects caused by two-dimensional vision analysis errors, and makes full use of the temporal relationship between video frames to improve the accuracy and real-time of 3D pose inference results of video stream. The method includes, for n(n >= 2) th frame of video, (1) inputting two-dimensional image of current frame and using shallow neural network module to generate image shallow map; 2)inputting a two-dimensional joint thermodynamic map of the human body generated by the (n-1)th frame and an image shallow map generated by the current frame to an LSTM module to generate a deep-levelcharacteristic map; 3) outputting the deep image characteristic map generated by the current frame to the residual module to generate a two-dimensional human body joint thermodynamic map of the current frame; 4) outputting the human body two-dimensional joint thermodynamic map of the current frame to a three-dimensional joint inference module for carrying out two-dimensional to three-dimensionalspatial mapping; and superimposing the three-dimensional human body joint thermodynamic map generated in each frame to generate a video stream of three-dimensional human body posture estimation.

Description

technical field [0001] The invention relates to a method for estimating three-dimensional human body poses for two-dimensional image and video streams, and belongs to the technical field of virtual reality. Background technique [0002] The 3D pose estimation of the human body is to accurately estimate the 3D positions of several joints of the human body (such as the head, shoulder, elbow, etc.). Due to the loss of depth information, estimating the position of 3D joint points of the human body from a 2D RGB video stream is a great challenge in the field of computer vision. [0003] With the development of Deep Convolutional Networks (Deep Convolutional Networks), more and more technological innovations focus on 3D human skeleton detection based on end-to-end deep neural networks. The more common 3D human pose estimation methods currently have the following two technical routes: [0004] Two-stage 3D joint point inference, as attached figure 1 As shown, the method is divid...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
IPC IPC(8): G06K9/00G06T17/00
CPCG06T17/00G06V20/647G06V40/10
Inventor 李帅胡韬于洋付延生
Owner QINGDAO RES INST OF BEIHANG UNIV
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products