Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Weak supervision time sequence behavior positioning method

A behavior positioning and weak supervision technology, applied in the field of computer vision, can solve problems such as being unable to be correctly identified, blurring video clips, etc., and achieve the effect of improving positioning performance and suppressing activation

Pending Publication Date: 2021-11-16
HUAIBEI NORMAL UNIVERSITY
View PDF0 Cites 1 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

But a large number of videos created in real life, in addition to the most recognizable video clips, there are also a large number of blurred video clips that cannot be correctly identified

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Weak supervision time sequence behavior positioning method
  • Weak supervision time sequence behavior positioning method
  • Weak supervision time sequence behavior positioning method

Examples

Experimental program
Comparison scheme
Effect test

Embodiment

[0052] A weakly supervised time-series behavior positioning method, the model diagram is as follows figure 1 shown, including the following steps:

[0053]Extract the original features of the video clips in the training set: extract the RGB and Flow frames of the video, divide them into non-overlapping 16-frame clips, obtain the RGB and Flow features of each clip through the feature extractor, and connect these two special features as The original characteristics of the video.

[0054] Build a feature embedding module, which is a neural network, including a convolutional layer and a Relu layer, so as to obtain the same embedded features as the original feature latitude;

[0055] Build a classification branch and introduce a non-action class; input the embedded features to the classification branch to obtain the initial category activation sequence;

[0056] According to the attention mechanism, the action attention module and the background attention module are respectively ...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention provides a weak supervision time sequence behavior positioning method. The method comprises the following steps: acquiring an embedded feature of a video clip; classifying the embedded feature, introducing a non-action class, and obtaining an initial class activation sequence corresponding to the embedded feature; constructing an action attention module and a background attention module according to an attention mechanism, constructing a context attention module by using the difference between the action attention module and the background attention module, and weighting the initial class activation sequence based on the three attention values to obtain an action instance class activation sequence, a background class activation sequence and a context class activation sequence; executing a top-k policy on the action instance activation sequence to obtain a video level classification score, wherein the action class of which the score is greater than a threshold value is the action class involved in a to-be-positioned video; and for the involved class, applying a threshold policy to corresponding action class activation sequence and action instance attention, so that behavior positioning is realized. According to the method, the attention model is concisely and effectively defined, action context and background activation is effectively inhibited, and the positioning performance is improved.

Description

technical field [0001] The invention relates to the technical field of computer vision, in particular to a weakly supervised temporal sequence behavior positioning method. Background technique [0002] In video applications such as video description, video summarization, video understanding, and intelligent surveillance, temporal behavior localization is one of the essential basic tasks. Its goal is to accurately locate the temporal boundaries of all action instances in an uncropped video and determine their action categories. However, an uncropped video usually has a large number of backgrounds and even multiple temporal actions, which poses a huge challenge for temporal behavior location. Due to the rapid development of deep learning, many methods have been proposed and achieved excellent performance. However, most of these methods are implemented under the definition of full supervision, which requires a large collection of videos with precise annotations during trainin...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
IPC IPC(8): G06K9/00G06K9/62G06N3/08
CPCG06N3/08G06F18/2431
Inventor 高向军刘梦雪葛方振刘怀愚李想沈龙凤洪留荣
Owner HUAIBEI NORMAL UNIVERSITY
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products