Unlock instant, AI-driven research and patent intelligence for your innovation.

Behavior classification method based on twin three-dimensional convolutional neural network

A three-dimensional convolution and neural network technology, applied in the field of video speech understanding, can solve the problems of ignoring video time characteristics, unsatisfactory classification algorithm effect, and inability to apply calculation speed.

Active Publication Date: 2020-04-17
TIANJIN UNIV
View PDF4 Cites 0 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

At present, there is no optimal solution to the video classification problem. One type of algorithm only uses the spatial characteristics of the video, ignoring the temporal characteristics of the video, which makes the effect of the classification algorithm unsatisfactory; the other type of algorithm uses both time and space. feature, this type of method describes the temporal characteristics of the video through the optical flow field, and abstracts the temporal characteristics for behavior classification, but the calculation speed of the optical flow field cannot be applied to the scene of real-time behavior classification, and the optical flow field does not describe the temporal characteristics of the video The best method, the performance of the behavior classification algorithm based on the optical flow field still has a lot of room for improvement

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Behavior classification method based on twin three-dimensional convolutional neural network
  • Behavior classification method based on twin three-dimensional convolutional neural network
  • Behavior classification method based on twin three-dimensional convolutional neural network

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0020] The technical solutions of the present invention will be described in detail below with reference to the accompanying drawings and embodiments.

[0021] like figure 1 As shown, it is a schematic diagram of the framework of the Siamese three-dimensional convolutional neural network on which the present invention is based. Including the abstract temporal feature branch network and the abstract spatial feature branch network, the structures of the two branches are the same, and the convolution kernels used are all three-dimensional convolution kernels. The abstract temporal features are passed through a deconvolutional network to generate an optical flow field. The splicing of abstract features is end-to-end, i.e. where f cat represents the feature after concatenation, f s represents the abstract spatial feature, f t represents an abstract temporal feature, and f s , The classifier is composed of a fully connected layer, and the output dimension of the fully conne...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses a behavior classification method based on a twin three-dimensional convolutional neural network. The method comprises the following steps: 1, adjusting the length and width ofan image frame to the size required by the twin three-dimensional convolutional neural network; 2, grouping an image frame sequence in groups of continuous 16 frames, inputting each group of image frames into the twin three-dimensional convolutional neural network, extracting abstract time features and abstract space features, and inputting the abstract time features into a deconvolution network to obtain an optical flow field; 3, calculating to obtain a total loss function L, wherein the expression is as follows: L = Lcls + alpha Lflow; and 4, optimizing network parameters by using a back propagation technology, so that the network performance is optimal. The algorithm provided by the invention has advantages in both speed and accuracy.

Description

technical field [0001] The present invention relates to the neighborhood of video speech understanding, in particular to a method for classifying behaviors in videos. Background technique [0002] Behavior classification is an important branch in the field of video understanding. Higher-level tasks such as generating video collections and describing video content in text are all based on behavior classification. Behavior classification is a technology that extracts the abstract semantic information of the video and judges the action category contained in the video according to the semantic information. At present, there is no optimal solution to the video classification problem. One type of algorithm only uses the spatial characteristics of the video, ignoring the temporal characteristics of the video, which makes the effect of the classification algorithm unsatisfactory; the other type of algorithm uses both time and space. This kind of method describes the temporal featur...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(China)
IPC IPC(8): G06K9/00G06N3/04G06N3/08
CPCG06N3/084G06V20/41G06N3/045
Inventor 周圆李鸿儒李绰李孜孜杨晶
Owner TIANJIN UNIV