Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Action detection model based on convolutional neural network

A convolutional neural network and motion detection technology, applied in the field of computer vision research, can solve problems such as low time efficiency and time-consuming motion positioning

Inactive Publication Date: 2017-05-10
BEIJING UNIV OF TECH
View PDF3 Cites 51 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

In the past, many motion detection methods used the sliding window method (sliding window). However, motion positioning based on the sliding window method is extremely time-consuming and time-efficient. In order to speed up motion positioning, Oneatra et al. proposed a Fisher Vector strategy that is close to normalization. This strategy uses a more efficient method than sliding windows, the branch-and-bound search algorithm

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Action detection model based on convolutional neural network
  • Action detection model based on convolutional neural network
  • Action detection model based on convolutional neural network

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0019] The realization and verification of the idea of ​​the action detection model in the present invention uses GPU (K80) as the computing platform, adopts CUDA as the GPU parallel computing framework, and selects Caffe as the CNN framework. The specific implementation steps are as follows:

[0020] Step 1: Preprocessing of video data

[0021] The video data required by this method needs to be split and saved in the form of "one frame, one picture", and the size of each frame of pictures must be consistent. There are currently many open video datasets to choose from, choose one or more according to the specific task. Secondly, it is necessary to calculate the optical flow for each frame in the data set, obtain the optical flow map corresponding to each frame of the picture, organize and save the optical flow map data set.

[0022] Step 2: Training of Faster RCNN

[0023]The two-channel Faster RCNN network is trained with the frame image data set and the optical flow image...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses an action detection model based on a convolutional neural network, and belongs to the field of computer visual studies. An efficient action detection model is constructed by using the convolutional neural network in deep learning, thereby recognizing an action from video and detecting and positioning the action. The action detection model is composed of a Faster RCNN (Regional Convolutional Neural Network), an SVM (Support Vector Machine) classifier and an action pipeline. Each part of the action detection model respectively completes the corresponding operation. The Faster RCNN acquires a plurality of regions of interest from each frame of picture, and extracts a feature from each region of interest. The detection model extracts the features by adopting a double-channel model, namely a Faster RCNN channel based on a frame picture and a Faster RCNN channel based on an optical flow picture, which are respectively used for extracting an appearance feature and an action feature. Then, the two features are fused into a time-space domain feature, the time-space domain feature is input to the SVM classifier, and an action type prediction value of the corresponding region is given by SVM classification. Finally, a final action detection result is given by the action pipeline from the perspective of video.

Description

technical field [0001] The invention belongs to the field of computer vision research, and constructs an efficient action detection model by using the method of convolutional neural network in deep learning, realizes action recognition from video and further detects and locates the action. Background technique [0002] Video recognition in the field of computer vision is divided into action classification and action detection. The problem to be solved by action classification is similar to "Is there an action or behavior like 'running' in this video?" Which frame sequence set the action appears in and where the action is in each frame". [0003] In recent years, benefiting from the great progress made in image recognition, video recognition has also made great progress. Most of the action recognition methods are proposed for action classification tasks. In fact, these methods can also be used for action detection tasks. J.Aggarwal, M.Ryoo, R.Poppe, etc. have summarized an...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(China)
IPC IPC(8): G06K9/00G06K9/62G06N3/04G06N3/08
CPCG06N3/08G06V20/42G06N3/045G06F18/2411G06F18/214
Inventor 刘波贾川川
Owner BEIJING UNIV OF TECH
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products