Multi-target unmarked attitude estimation method based on deep convolutional neural network

A pose estimation and neural network technology, applied in the field of computer vision, can solve problems such as difficulty in the use of personnel, camera distortion, uneven lighting, etc., and achieve the effects of less loss of prediction accuracy, improved accuracy, and good robustness

Active Publication Date: 2019-09-06
NANJING UNIV OF SCI & TECH
View PDF2 Cites 23 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

However, the research on pose estimation faces many challenges, such as the number of targets in an image is usually not predetermined, and it is easy for targets to occlude each other in the case of multiple targets.
Moreover, as the number of targets increases, the operating efficiency of the algorithm is also facing huge challenges.
Before deep learning was widely used, the graph structure (Pictorial Structures) model was widely used in pose estimation, but the results based on the graph structure model were not very accurate, and it was difficult to extend to multi-target pose estimation
[0003] The proposal of deep learning algorithm provides a new direction for multi-target pose estimation. At present, there are two main ideas for multi-target pose estimation. One is ...

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Multi-target unmarked attitude estimation method based on deep convolutional neural network
  • Multi-target unmarked attitude estimation method based on deep convolutional neural network
  • Multi-target unmarked attitude estimation method based on deep convolutional neural network

Examples

Experimental program
Comparison scheme
Effect test

Embodiment

[0054] The multi-target unmarked attitude estimation method based on deep convolutional neural network of the present invention comprises the following contents:

[0055] 1. Acquire image sequences containing target behaviors from videos. Specifically:

[0056] Using the clustering method of visual appearance (K-means), collect image sequences of the behavior of the target of interest under different brightness conditions and background conditions, and ensure that the training data set contains a sufficient number of image sequences (100-200). The image sequence collected in this embodiment is as follows figure 2 shown.

[0057] 2. For each image in the collected image sequence, manually mark the position and category of each target feature part in the same order, and construct a training data set and a test data set according to the marked image sequence. Specifically:

[0058] 70% of the image sequences are randomly selected as the training image set, and the remaining ...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention discloses a multi-target unmarked attitude estimation method based on a deep convolutional neural network. The multi-target unmarked attitude estimation method comprises the following steps: acquiring an image sequence containing a target behavior from a video; for each image, marking the position and category of each target feature part, and constructing a training data set and a test data set; constructing a multi-target attitude estimation model; training the multi-target attitude estimation model to adjust the weight of the multi-target attitude estimation model, and optimizing the multi-target attitude estimation model; and carrying out precision evaluation on the multi-target attitude estimation model, carrying out target attitude estimation or retraining according to arelation between precision and an expected value, and optimizing the multi-target attitude estimation model. According to the multi-target unmarked attitude estimation method, the accuracy can reachthe human level only through a small number of training images; the target feature part can be effectively tracked under the conditions of messy background, non-uniform illumination, camera distortionand the like; and a good effect can be achieved for different types of targets.

Description

technical field [0001] The invention belongs to the field of computer vision, in particular to a multi-target unmarked pose estimation method based on a deep convolutional neural network. Background technique [0002] The problem of multi-target pose estimation based on computer vision refers to estimating the position and associated information of various parts based on image features. It has a wide range of applications and strong practical value in the fields of human-computer interaction, video surveillance, virtual reality, and automatic driving. However, the research on pose estimation faces many challenges, such as the number of objects in an image cannot be determined in advance, and it is easy for objects to occlude each other in the case of multiple objects. Moreover, as the number of targets increases, the operating efficiency of the algorithm is also facing huge challenges. Before deep learning was widely used, the graph structure (Pictorial Structures) model wa...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
IPC IPC(8): G06K9/00G06K9/62G06N3/04G06N3/08
CPCG06N3/084G06V40/20G06V20/40G06V2201/07G06N3/045G06F18/214G06F18/241Y02T10/40
Inventor 白宏阳周育新李政茂郑浦徐啸康郭宏伟梁华驹
Owner NANJING UNIV OF SCI & TECH
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products