Action recognition method based on attention mechanism of convolution recurrent neural network

A technology of recursive neural network and convolutional neural network, which is applied in the field of computer vision action recognition, can solve problems such as the inability to effectively extract salient areas, and achieve the effect of improving accuracy

Active Publication Date: 2017-10-20
DALIAN UNIV OF TECH
View PDF3 Cites 74 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0007] Aiming at the problem that the salient region cannot be effectively extracted in the process of action recognition, the present invention proposes an action recognition method based on a convolutional recurrent neural network ba

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Action recognition method based on attention mechanism of convolution recurrent neural network
  • Action recognition method based on attention mechanism of convolution recurrent neural network
  • Action recognition method based on attention mechanism of convolution recurrent neural network

Examples

Experimental program
Comparison scheme
Effect test

Example Embodiment

[0030] An embodiment of the present invention provides an action recognition method based on an attention mechanism. The specific embodiments discussed are merely illustrative of implementations of the invention, and do not limit the scope of the invention. Embodiments of the present invention will be described in detail below in conjunction with the accompanying drawings, specifically including the following steps:

[0031] 1 Data preprocessing. The size of the RGB image of the original video frame is not uniform, which is not suitable for subsequent processing. The present invention cuts the original image so that its size can be unified. At the same time, in order to speed up the subsequent processing, the present invention performs normalization processing on the image.

[0032] 2 feature extraction. In view of the success of the GoogleNet neural network in image feature representation, the present invention regards a video as an image collection composed of multiple fr...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The present invention belongs to the field of computer vision action recognition, and proposes an action recognition method based on the attention mechanism of the convolution recurrent neural network, in order to solve the problem that the obvious region cannot be effectively extracted in the action recognition and to improve the accuracy of classification. The method comprises: using the convolution neural network to extract the feature of the action video automatically; using the spatial transformation network to realize the attention mechanism based on the feature map, and extracting the obvious feature region by using the attention mechanism to generate the target feature map; and inputting the target feature map into the convolutional recurrent neural network to produce the final action recognition result. Experiments show that the proposed method has achieved good results on the benchmark action video test set such as UCF-11, HMDB-51, and the like, and improves the accuracy of action recognition.

Description

technical field [0001] The invention belongs to the field of computer vision action recognition, and relates to an action recognition method based on a convolution recursive neural network based on an attention mechanism. Background technique [0002] With the development of the Internet, video has become an indispensable part of today's big data, which promotes the research on video classification and produces a large number of novel technologies. Compared with images, video has a lot of rich and contextual information, which requires the ability to build a good model to capture the features contained in it. At present, the understanding of video content has become a problem to be solved in video processing. The method of deep learning subverts the design ideas of traditional algorithms in many fields such as speech recognition, image classification, and text understanding, and gradually forms an end-to-end model starting from training data. A new mode for outputting the ...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
IPC IPC(8): G06K9/00G06N3/04
CPCG06V40/23G06V20/41G06N3/045
Inventor 葛宏伟宇文浩闫泽航
Owner DALIAN UNIV OF TECH
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products