Multistage video action detecting method

A motion detection, multi-stage technology, applied in the direction of instruments, biological neural network models, character and pattern recognition, etc., can solve the problems that 3DCNN cannot learn motion features, increase the difficulty of network training, and break continuity, so as to improve motion detection results, avoid hard-to-learn effects, improve accuracy

Active Publication Date: 2018-10-16
UNIV OF SCI & TECH OF CHINA
View PDF6 Cites 22 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

However, since the candidate classification network requires the same time length of the video clips input by the network, this method obtains video clips of different time lengths by controlling the frequency of downsampling. However, using the same network structure to train video clips obtained with different sampling frequencies will lead to The difference within the class increases, and the continuity within the action is destroyed, so that the 3D CNN cannot learn good motion features and increases the difficulty of network training.
Moreover, this method only classifies the action categories of the video clips, and does not realize the fine-tuning of the boundaries of the video clips, so that the improvement of the accuracy of action positioning is limited.

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Multistage video action detecting method
  • Multistage video action detecting method
  • Multistage video action detecting method

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0021] The technical solutions in the embodiments of the present invention will be clearly and completely described below in conjunction with the accompanying drawings in the embodiments of the present invention. Obviously, the described embodiments are only some of the embodiments of the present invention, not all of them. Based on the embodiments of the present invention, all other embodiments obtained by persons of ordinary skill in the art without making creative efforts belong to the protection scope of the present invention.

[0022] Embodiments of the present invention provide a multi-stage video motion detection method, such as figure 1 As shown, it mainly includes the following three steps:

[0023] 1. For the input uncut video, through the binary classification and voting fusion strategy based on the deep residual network, generate coarse motion clips that fuse multi-scale sampling and single-scale training.

[0024] Such as figure 2 As shown, in order to generate...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention discloses a multistage video action detecting method. The multistage video action detecting method comprises the steps that rough action fragments fusing multiscale sampling and single-scale training are generated by an input video which is not clipped through a dichotomy and voting fusing strategy based on a deep residual network; the rough action fragments are subjected to action classification and action boundary joint distinguishing through a statistical fusion strategy based on a frame-level action recognition result, and primary action detecting fragments are obtained; in combination of IoU between the primary action detecting fragments, an improved non-maximum suppression algorithm is utilized for processing the primary action detecting fragments, and finally the action detecting result, namely, the action classification and beginning and ending time position of each video action detecting fragment, of the video which is not clipped is obtained. According to the method, the action classification accuracy and action locating precision can be improved.

Description

technical field [0001] The invention relates to the technical field of video motion detection, in particular to a multi-stage video motion detection method. Background technique [0002] With the rapid development of network and multimedia technology, video has become an important carrier for people to obtain information, and the number of videos is growing explosively, so the analysis and understanding technology of video content is very important. Uncropped videos usually contain multiple action instance segments and a large number of irrelevant background segments, where the action occurs, the time interval, and the action category label are unknown. The video action detection task is to be able to identify the category labels of multiple action instances in the uncut video and locate the start time and end time of the action instance. As one of the important research directions of current video processing technology, this task is widely used in intelligent Surveillance,...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
Patent Type & Authority Applications(China)
IPC IPC(8): G06K9/00G06N3/04
CPCG06V20/42G06N3/045
Inventor 王子磊赵琰
Owner UNIV OF SCI & TECH OF CHINA
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products