Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Dual-module neural network structure video object segmentation method

A video object and network structure technology, applied in the field of computer vision, can solve problems such as complex background, occlusion, and inability to achieve efficient segmentation, and achieve the effects of enhancing discrimination, suppressing noise influence, and saving costs

Pending Publication Date: 2020-03-24
ANHUI UNIVERSITY
View PDF4 Cites 7 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

These methods usually use the first frame for data enhancement, and model adaptation relies heavily on the fine-tuning model. In the video, complex backgrounds, occlusions or fast movements, and camera shake oscillations cannot achieve efficient segmentation problems.

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Dual-module neural network structure video object segmentation method
  • Dual-module neural network structure video object segmentation method
  • Dual-module neural network structure video object segmentation method

Examples

Experimental program
Comparison scheme
Effect test

Embodiment

[0105] The experimental hardware environment of the present invention is: 3.4GHz Intel(R) Core(TM) i5-7500 CPU and GTX 1080TiGPU PC, 16 memory, Ubuntu18.04 operating system, based on the open source framework Pytorch depth framework. An image size of 854x480 is used for training and testing. test results (such as Figure 4 Figure 5 ) data set comes from DAVIS public video image segmentation data set.

[0106] First, for the given first frame and the mask of the first frame (such as figure 2 shown in 1 and 2). Image pairs from 1 to 100 are generated by transforming the network ( figure 2 shown in 4). Select candidate regions of interest through the target proposal box ( figure 2 shown in 5). After adding the tracker in the area of ​​interest, input it into the RoISeg network for training ( figure 2 shown in 6). The output feature map from the last convolutional layer in the RoISeg network ( figure 2 Shown in 7) are respectively input into the spatial attention mo...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention provides a dual-module neural network structure video object segmentation method, for solving the problem that a video object segmentation result is not ideal due to noise interference in a video object segmentation process. The dual-module neural network structure video object segmentation method comprises the following steps: inputting a first frame image and a mask of a first frame into a transformation network to generate an image pair; generating a target proposal box for each image pair to determine whether the image pair is a region of interest; adding a tracker into the region of interest, inputting the tracker into a segmentation network of interest to train a learning model, and outputting the learning model; outputting a feature map from the last layer of convolution of the interested segmentation network and respectively inputting into a space attention module and a channel attention module; and finally, fusing the feature maps output by the two attention modules, and outputting a final segmentation mask result through convolution layer operation. According to the dual-module neural network structure video object segmentation method, a good segmentation experiment result is obtained on the DAVIS video data set.

Description

technical field [0001] The invention is in the field of computer vision, especially relates to video object segmentation processing with large-scale changes in video and inaccurate dynamic appearance changes, and specifically a method for video object segmentation with a dual-module neural network structure. Background technique [0002] In recent years, with the rapid development of computer vision technology, convolutional neural network in deep learning has received great attention in various research fields, and video object segmentation technology has become an important content that researchers have paid attention to in recent years. Video segmentation technology is increasingly showing its important position. Its applications in scene understanding, video labeling, driverless cars and object detection have all been rapidly developed in video segmentation technology. It can be said that the advancement of video segmentation technology drives the overall development of...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
IPC IPC(8): G06T7/10G06T7/207
CPCG06T7/10G06T7/207G06T2207/10016G06T2207/20081G06T2207/20084G06T2207/20104G06T2207/20221Y02T10/40
Inventor 汪粼波陈彬彬方贤勇
Owner ANHUI UNIVERSITY
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products