Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Action recognition method based on double-flow space-time attention mechanism

An action recognition and attention technology, applied in the fields of video classification and computer vision, which can solve the problems of complex temporal attention network structure, low recognition efficiency, and inaccurate regions.

Active Publication Date: 2020-09-04
SHENYANG INST OF ENG
View PDF10 Cites 22 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0004] (1) When extracting motion salient spatial region information, only one spatial attention network is used to focus on multiple salient regions of the frame, resulting in inaccurate extraction of some regions;
[0005] (2) The temporal attention network designed using LSTM has a complex structure and must process video frames in chronological order, and the recognition efficiency is low

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Action recognition method based on double-flow space-time attention mechanism
  • Action recognition method based on double-flow space-time attention mechanism
  • Action recognition method based on double-flow space-time attention mechanism

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0044] The technical solution of the present invention will be further described in detail below in conjunction with the accompanying drawings.

[0045] The present invention designs an action recognition method T-STAM based on a dual-stream spatio-temporal attention mechanism, see figure 1 , the method includes the following steps.

[0046]S1: Process the video and select RGB frames, and obtain the optical flow diagram of the selected RGB frames;

[0047] S2: The channel attention mechanism can learn the importance of each feature channel, improve the channel features that are useful for current recognition according to the importance, and suppress the channel features with weak recognition to obtain the structure. Therefore, the present invention introduces the channel attention network SE-Net to the dual-stream basic network BN-Inception to obtain SE-BN-Inception that can model channel features. The channel attention network SE-Net is introduced into the dual-stream basic...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses an action recognition method T-STAM based on a double-flow space-time attention mechanism. The method comprises the following steps: S1, processing a video to obtain an opticalflow graph of an RGB frame; s2, fusing the channel attention network SE-Net into a double-flow basic network BN-Inception to obtain an SE-BN-Inception; s3, inputting the selected RGB frame and optical flow field information into SE-BN-Inception, and modeling the dependency relationship of different channels in the features to obtain a feature vector X of the video; s4, inputting the feature X into a CNN-based time attention network to calculate a time attention score corresponding to each frame; s5, inputting the feature X into a multi-space attention network, and extracting a plurality of motion space salient regions of the frame; s6, fusing the spatial and temporal features to further enhance the feature expression of the video, and fusing the two streams according to different weightsto obtain an action recognition result.

Description

technical field [0001] The invention relates to the fields of computer vision, video classification and the like, and in particular provides an action recognition method T-STAM based on a dual-stream spatio-temporal attention mechanism. Background technique [0002] In recent years, with the rise of deep learning, methods based on convolutional neural networks have been widely used in the field of video action recognition research. Among them, the dual-stream method inputs RGB into the CNN to obtain appearance information, and inputs multi-frame optical flow fields into the CNN to obtain motion information, which can effectively combine the spatiotemporal information in the video and is relatively better in performance. However, the two-stream method ignores the connection of different channel information when extracting video features. In addition, it processes the sampled frames in the video equally, does not distinguish the information of different positions of the frame...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(China)
IPC IPC(8): G06T7/269G06K9/00G06K9/62
CPCG06T7/269G06T2207/10016G06T2207/20081G06T2207/20084G06V40/20G06F18/24
Inventor 代钦王黎明李怡颖王洪江刘芳
Owner SHENYANG INST OF ENG
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products