Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Multi-model fusion video hand division method based on Kinect

A technology that integrates video and multiple models. It is used in character and pattern recognition, instruments, computer parts, etc., and can solve problems such as depth information noise, lack of segmentation effect, and inaccurate edges.

Active Publication Date: 2013-07-31
SOUTH CHINA UNIV OF TECH
View PDF4 Cites 30 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

However, the depth information provided by Kinect has defects such as noise and inaccurate edges.
Therefore, simply using depth information often does not get good segmentation results. Some researchers combine skin color model and depth model to build a more robust segmentation model. However, due to Kinect's own imaging principle, these two information are out of sync. Therefore, there is a certain delay between the color image and the depth image, and how to register and make up for the non-overlapping parts has become a new problem

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Multi-model fusion video hand division method based on Kinect
  • Multi-model fusion video hand division method based on Kinect
  • Multi-model fusion video hand division method based on Kinect

Examples

Experimental program
Comparison scheme
Effect test

Embodiment

[0061] Such as figure 1 As shown, it is a block diagram of the system structure of the present invention. After the user video is obtained through Kinect, the depth, skin color and background models are used to segment respectively, and the overlapping ratios of the segmentation results of the three models are calculated, and input to a neural network evaluation system . The neural network outputs the "confidence coefficient" of each of the three models. The segmentation results of each model are weighted by the confidence coefficient and then linearly accumulated. Finally, the final segmentation result is obtained through a threshold function, and the non-human hand area is extracted from the segmentation result to update the background model regularly. .

[0062] The structure of the neural network evaluation system is as follows: figure 2 As shown, the input layer 1 accepts the pairwise overlap rate of the three model segmentation results as input, that is, OLR s_d ,OLR...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention provides a multi-model fusion video hand division method based on Kinect, which comprises the following steps of (1) capturing video information, (2) dividing images in a video respectively to obtain division results, namely a depth model, a skin color model and a background model, (3) calculating an overlapping rate of every two division results as a characteristic of judging division effects of the results and inputting the three overlapping rates into a neural network, (4) allowing the neural network to output three coefficients (namely confidence coefficients) showing respective reliability of the three models, and weighting the three division results with the confidence coefficients, (5) conducting linear superposition on the weighted division results of the three models, (6) outputting a final binary image of a superposed result through a threshold function and finally dividing an obtained video hand region, and (7) updating the background model, wherein the division results are expressed as binary images. The method has the advantages of low cost, good flexibility and the like.

Description

technical field [0001] The invention relates to a computer image processing and pattern recognition technology, in particular to a Kinect-based multi-model fusion video hand segmentation method. Background technique [0002] Video hand segmentation is an important field of computer vision and artificial intelligence. As a basic step in a series of human-computer interaction applications such as gesture remote control, virtual writing and painting, sign language recognition, etc., it has an important impact on subsequent steps. Traditional video hand segmentation methods can be divided into two main areas: segmentation methods based on 2D vision and segmentation methods based on 3D hand modeling. Previous researchers have their own limitations in the attempts of 2D vision methods. sex. This includes the widely used skin tone model, but it is susceptible to lighting changes and cannot resolve the overlap between human hands and skin-like regions. On the other hand, the frame...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(China)
IPC IPC(8): G06K9/46G06K9/62G06K9/00
Inventor 金连文叶植超张鑫
Owner SOUTH CHINA UNIV OF TECH
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products