Video target segmentation method based on motion attention

A target segmentation and attention technology, applied in the fields of image processing and computer vision, can solve the problems of inability to obtain the precise position of the target object, limited motion mode of the segmentation result, and drift of the target object, so as to reduce useless features and improve robustness. , Split effect for precise effect

Active Publication Date: 2020-05-15
BEIJING UNIV OF TECH
View PDF7 Cites 8 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0005] The problem to be solved by the present invention is: in video object segmentation, only relying on the segmentation result of the previous frame to determine the target object in the current frame, the precise position of the target object cannot be obtained, and even due to excessive dependence on the segmentation result of the previous frame, resulting in The target object drifts; and the existing video target segmentation methods based on motion information, most of the target object segmentation is based on the optical flow information between the current frame and the previous frame, which not only has a large amount of calculation, but also limits the segmentation results specific sport mode

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Video target segmentation method based on motion attention
  • Video target segmentation method based on motion attention
  • Video target segmentation method based on motion attention

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0031] The present invention proposes a video target segmentation method based on motion attention. The method first obtains the feature maps of the first frame and the current frame, and then combines the first frame feature map, the current frame feature map and the previous frame motion attention network The position information of the target object predicted by the middle memory module is input to the motion attention network to obtain the segmentation result of the current frame. The invention is suitable for video target segmentation, has good robustness and accurate segmentation effect.

[0032] The present invention will be described in more detail below in conjunction with specific examples and drawings.

[0033] The present invention includes the following steps:

[0034] 1) Obtain the YouTube and Davis data sets as the training set and test set of the model respectively;

[0035] 2) Preprocessing the training data. Cut each training sample (video frame) and the first fram...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention provides a video target segmentation method based on motion attention, and the method comprises the steps: adding a channel feature map outputted by a channel attention module and a position feature map outputted by a motion attention module, and obtaining a segmentation result of a current frame, wherein the input of the channel attention module is the feature map Ft of the currentframe and an appearance feature map F0 of a target object provided by the first frame; enabling the channel attention module to calculate the association between the input feature map Ft and the F0 channel, wherein the output channel feature map reflects the object with the appearance closest to the target object in the current frame, the input of the motion attention module is the current frame feature map Ft and the position information Ht-1 of the target object predicted by the memory module in the previous frame motion attention network; enabling the motion attention module to calculate the association between the positions of the input feature map Ft and the Ht-1, wherein the output position feature map reflects the approximate position of the target object in the current frame. According to the invention, two factors of appearance and position are combined to realize more accurate segmentation of the video target.

Description

Technical field [0001] The invention belongs to the field of image processing and computer vision, and relates to a method for segmenting video objects, and in particular to a method for segmenting video objects based on motion attention. Background technique [0002] Video object segmentation is a prerequisite for solving many video tasks, and it plays an important role in object recognition, video compression and other fields. Video target segmentation can be defined as tracking the target and segmenting the target object according to the target mask. According to whether there is an initial mask, video target segmentation can be divided into semi-supervised and unsupervised methods. The semi-supervised segmentation method is to manually initialize the segmentation mask in the first frame of the video, and then track and segment the target object. The unsupervised method automatically segment the target object in a given video based on a certain mechanism without any prior inf...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
Patent Type & Authority Applications(China)
IPC IPC(8): G06T7/215G06T5/30
CPCG06T7/215G06T5/30G06T2207/10016G06T2207/20081G06T2207/20084
Inventor 付利华杨寒雪杜宇斌姜涵煦
Owner BEIJING UNIV OF TECH
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products